[GitHub] [flink] flinkbot commented on issue #11405: [FLINK-16413]Reduce hive source parallelism when limit push down

2020-03-13 Thread GitBox
flinkbot commented on issue #11405: [FLINK-16413]Reduce hive source parallelism 
when limit push down
URL: https://github.com/apache/flink/pull/11405#issuecomment-599020246
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit b196447b93a3c1bdb64fce172c2aea305e16093b (Sat Mar 14 
06:54:16 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhangjun888 opened a new pull request #11405: [FLINK-16413]Reduce hive source parallelism when limit push down

2020-03-13 Thread GitBox
zhangjun888 opened a new pull request #11405: [FLINK-16413]Reduce hive source 
parallelism when limit push down
URL: https://github.com/apache/flink/pull/11405
 
 
   
   
   ## What is the purpose of the change
   
   Reduce hive source parallelism when limit push down
   
   
   ## Brief change log
   
   when limit push down ,set the parallelism to min(parallelism,limit)
   
   
   ## Verifying this change
   
   *(Please pick either of the following options)*
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   *(or)*
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   *(or)*
   
   This change added tests and can be verified as follows:
   
   *(example:)*
 - *Added integration tests for end-to-end deployment with large payloads 
(100MB)*
 - *Extended integration test for recovery after master (JobManager) 
failure*
 - *Added test that validates that TaskInfo is transferred only once across 
recoveries*
 - *Manually verified the change by running a 4 node cluser with 2 
JobManagers and 4 TaskManagers, a stateful streaming program, and killing one 
JobManager and two TaskManagers during the execution, verifying that recovery 
happens correctly.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector:no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16254) Support -p/--parallelism option for StatefulFunctionsClusterEntryPoint

2020-03-13 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-16254:
---
Labels: pull-request-available  (was: )

> Support -p/--parallelism option for StatefulFunctionsClusterEntryPoint
> --
>
> Key: FLINK-16254
> URL: https://issues.apache.org/jira/browse/FLINK-16254
> Project: Flink
>  Issue Type: New Feature
>  Components: Stateful Functions
>Affects Versions: statefun-1.1
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Major
>  Labels: pull-request-available
>
> Currently the only way for users to specify parallelism > 1 for Stateful 
> Functions applications is to provide a value for {{parallelism.default}} via 
> {{flink-conf.yaml}}.
> That is not so nice to use, as users would essentially need to rebuild the 
> Stateful Functions application image just to change the parallelism.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-statefun] tzulitai opened a new pull request #58: [FLINK-16254] Support -p / --parallelism command option

2020-03-13 Thread GitBox
tzulitai opened a new pull request #58: [FLINK-16254] Support -p / 
--parallelism command option
URL: https://github.com/apache/flink-statefun/pull/58
 
 
   This PR adds support for specifying parallelism for Stateful Function 
applications using the `-p / --parallelism` command.
   
   The new behaviour as as follows:
   - If `-p` is defined in the command line, then that value is always used as 
the parallelism.
   - Otherwise, the `parallelism.default` value in `flink-conf.yaml` is used, 
which by default is 1 if not present.
   
   ---
   
   ### Changelog
   
   - 27e7abc Adds parsing logic of the parallelism option from the command 
line, and adds the property to `StatefulFunctionsClusterConfiguration`.
   - 07c8dd2 Resolves the parallelism to use in 
`StatefulFunctionsJobGraphRetriever`, and uses the resolved value to create the 
JobGraph.
   - ee61047 Adapt the E2E tests, so that they use the `-p` command to specify 
parallelism.
   
   ---
   
   ### Verfying
   
   The changes to the E2E tests in ee61047 verifies this change: `mvn clean 
verify -Prun-e2e-tests`


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16591) Flink-zh Doc show a wrong Email address

2020-03-13 Thread Zili Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17059181#comment-17059181
 ] 

Zili Chen commented on FLINK-16591:
---

Yes of course! Please mark the issue as "In Progress" by clicking "Start 
Progress" button above when you start progress.

> Flink-zh Doc show a wrong Email address
> ---
>
> Key: FLINK-16591
> URL: https://issues.apache.org/jira/browse/FLINK-16591
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: forideal
>Assignee: forideal
>Priority: Minor
>
> The Link is https://flink.apache.org/zh/community.html
> 发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, 
> user, user-zh 等等)
> The right Email pattern is listname-subscr...@flink.apache.org.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-16591) Flink-zh Doc show a wrong Email address

2020-03-13 Thread Zili Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zili Chen reassigned FLINK-16591:
-

Assignee: forideal

> Flink-zh Doc show a wrong Email address
> ---
>
> Key: FLINK-16591
> URL: https://issues.apache.org/jira/browse/FLINK-16591
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: forideal
>Assignee: forideal
>Priority: Minor
>
> The Link is https://flink.apache.org/zh/community.html
> 发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, 
> user, user-zh 等等)
> The right Email pattern is listname-subscr...@flink.apache.org.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16592) The doc of Streaming File Sink has a mistake of grammar

2020-03-13 Thread Chen (Jira)
Chen created FLINK-16592:


 Summary: The doc of Streaming File Sink has a mistake of grammar
 Key: FLINK-16592
 URL: https://issues.apache.org/jira/browse/FLINK-16592
 Project: Flink
  Issue Type: Task
  Components: Documentation
Reporter: Chen
 Attachments: image-2020-03-14-12-04-27-065.png

In the following link, there is a mistake in grammar.

[https://ci.apache.org/projects/flink/flink-docs-release-1.10/zh/dev/connectors/streamfile_sink.html]

!image-2020-03-14-12-04-27-065.png|width=567,height=369!

it should be

{{}}
{code:java}
final StreamingFileSink sink = StreamingFileSink
   .forRowFormat(new Path(outputPath), new SimpleStringEncoder("UTF-8"))
   .withRollingPolicy(
  DefaultRollingPolicy.builder()
 .withRolloverInterval(TimeUnit.MINUTES.toMillis(15))
 .withInactivityInterval(TimeUnit.MINUTES.toMillis(5))
 .withMaxPartSize(1024 * 1024 * 1024)
 .build())
   .build();
{code}
 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16591) Flink-zh Doc show a wrong Email address

2020-03-13 Thread forideal (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

forideal updated FLINK-16591:
-
Description: 
The Link is https://flink.apache.org/zh/community.html
发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, user, 
user-zh 等等)

The right Email pattern is listname-subscr...@flink.apache.org.

  was:
发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, user, 
user-zh 等等)

The right Email pattern is listname-subscr...@flink.apache.org.


> Flink-zh Doc show a wrong Email address
> ---
>
> Key: FLINK-16591
> URL: https://issues.apache.org/jira/browse/FLINK-16591
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: forideal
>Priority: Minor
>
> The Link is https://flink.apache.org/zh/community.html
> 发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, 
> user, user-zh 等等)
> The right Email pattern is listname-subscr...@flink.apache.org.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16591) Flink-zh Doc show a wrong Email address

2020-03-13 Thread forideal (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17059175#comment-17059175
 ] 

forideal commented on FLINK-16591:
--

Hi [~tison] Can you assign this issue to me ? Thank you!

> Flink-zh Doc show a wrong Email address
> ---
>
> Key: FLINK-16591
> URL: https://issues.apache.org/jira/browse/FLINK-16591
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: forideal
>Priority: Minor
>
> 发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, 
> user, user-zh 等等)
> The right Email pattern is listname-subscr...@flink.apache.org.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16591) Flink-zh Doc‘s show a wrong Email address

2020-03-13 Thread forideal (Jira)
forideal created FLINK-16591:


 Summary: Flink-zh Doc‘s show a wrong Email address
 Key: FLINK-16591
 URL: https://issues.apache.org/jira/browse/FLINK-16591
 Project: Flink
  Issue Type: Bug
  Components: Documentation
Reporter: forideal


发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, user, 
user-zh 等等)

The right Email pattern is listname-subscr...@flink.apache.org.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16591) Flink-zh Doc show a wrong Email address

2020-03-13 Thread forideal (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

forideal updated FLINK-16591:
-
Summary: Flink-zh Doc show a wrong Email address  (was: Flink-zh Doc‘s show 
a wrong Email address)

> Flink-zh Doc show a wrong Email address
> ---
>
> Key: FLINK-16591
> URL: https://issues.apache.org/jira/browse/FLINK-16591
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: forideal
>Priority: Minor
>
> 发送一封不包含任何内容或主题的邮件到 subscribe-listn...@flink.apache.org(替换 listname 为 dev, 
> user, user-zh 等等)
> The right Email pattern is listname-subscr...@flink.apache.org.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16379) Introduce fromValues in TableEnvironment

2020-03-13 Thread Zhenghua Gao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16379?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17059150#comment-17059150
 ] 

Zhenghua Gao commented on FLINK-16379:
--

+1 to merge fromElements to fromValues. 

Construction empty table with Values may be useless because plan with empty 
LogicalValues could be optimized. 

> Introduce fromValues in TableEnvironment
> 
>
> Key: FLINK-16379
> URL: https://issues.apache.org/jira/browse/FLINK-16379
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Dawid Wysakowicz
>Assignee: Dawid Wysakowicz
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Introduce a fromValues method to TableEnvironment similar to {{VALUES}} 
> clause in SQL
> The suggested API could look like:
> {code}
>   /**
>* Creates a Table from a given row constructing expressions.
>*
>* Examples:
>*
>* You can use {@link Expressions#row(Object, Object...)} to create 
> a composite rows:
>* {@code
>*  tEnv.fromValues(
>*  row(1, "ABC"),
>*  row(2L, "ABCDE")
>*  )
>* }
>* will produce a Table with a schema as follows:
>* {@code
>*  root
>*  |-- f0: BIGINT NOT NULL
>*  |-- f1: VARCHAR(5) NOT NULL
>* }
>*
>* ROWs that are a result of e.g. a function call are not flattened
>* {@code
>*  public class RowFunction extends ScalarFunction {
>*  @DataTypeHint("ROW")
>*  Row eval();
>*  }
>*
>*  tEnv.fromValues(
>*  call(new RowFunction()),
>*  call(new RowFunction())
>*  )
>* }
>* will produce a Table with a schema as follows:
>* {@code
>*  root
>*  |-- f0: ROW<`f0` BIGINT, `f1` VARCHAR(5)>
>* }
>*
>* The row constructor can be dropped to create a table with a 
> single row:
>*
>* ROWs that are a result of e.g. a function call are not flattened
>* {@code
>*  tEnv.fromValues(
>*  1,
>*  2L,
>*  3
>*  )
>* }
>* will produce a Table with a schema as follows:
>* {@code
>*  root
>*  |-- f0: BIGINT NOT NULL
>* }
>*
>* @param expressions Expressions for constructing rows of the VALUES 
> table.
>*/
>   Table fromValues(Expression... expressions);
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16583) SQLClientKafkaITCase.testKafka failed with SqlClientException

2020-03-13 Thread Hequn Cheng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17059148#comment-17059148
 ] 

Hequn Cheng commented on FLINK-16583:
-

Another instance: https://api.travis-ci.org/v3/job/661966609/log.txt

> SQLClientKafkaITCase.testKafka failed with SqlClientException
> -
>
> Key: FLINK-16583
> URL: https://issues.apache.org/jira/browse/FLINK-16583
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Table SQL / Client, Tests
>Reporter: Zhijiang
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.11.0
>
>
> The end-to-end test {{SQLClientKafkaITCase.testKafka}} failed with
> {code:java}
> 18:13:02.425 [ERROR] testKafka[0: kafka-version:0.10 
> kafka-sql-version:.*kafka-0.10.jar](org.apache.flink.tests.util.kafka.SQLClientKafkaITCase)
>   Time elapsed: 32.246 s  <<< ERROR!
> java.io.IOException: 
> Process execution failed due error. Error output:Mar 12, 2020 6:11:46 PM 
> org.jline.utils.Log logr
> WARNING: Unable to create a system terminal, creating a dumb terminal (enable 
> debug logging for more information)
> Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
> Could not submit given SQL update statement to cluster.
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:131)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:178)
>   at 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase.insertIntoAvroTable(SQLClientKafkaITCase.java:178)
>   at 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase.testKafka(SQLClientKafkaITCase.java:151)
> 18:13:02.425 [ERROR] testKafka[1: kafka-version:0.11 
> kafka-sql-version:.*kafka-0.11.jar](org.apache.flink.tests.util.kafka.SQLClientKafkaITCase)
>   Time elapsed: 34.539 s  <<< ERROR!
> java.io.IOException: 
> Process execution failed due error. Error output:Mar 12, 2020 6:12:21 PM 
> org.jline.utils.Log logr
> WARNING: Unable to create a system terminal, creating a dumb terminal (enable 
> debug logging for more information)
> Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
> Could not submit given SQL update statement to cluster.
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:131)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:178)
>   at 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase.insertIntoAvroTable(SQLClientKafkaITCase.java:178)
>   at 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase.testKafka(SQLClientKafkaITCase.java:151)
> {code}
> [https://api.travis-ci.org/v3/job/661535183/log.txt]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zhangjun888 closed pull request #11398: [FLINK-16413]Reduce hive source parallelism when limit push down

2020-03-13 Thread GitBox
zhangjun888 closed pull request #11398: [FLINK-16413]Reduce hive source 
parallelism when limit push down
URL: https://github.com/apache/flink/pull/11398
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-16392) oneside sorted cache in intervaljoin

2020-03-13 Thread Chen Qin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16392?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Qin closed FLINK-16392.

Resolution: Feedback Received

> oneside sorted cache in intervaljoin
> 
>
> Key: FLINK-16392
> URL: https://issues.apache.org/jira/browse/FLINK-16392
> Project: Flink
>  Issue Type: Improvement
>  Components: API / DataStream
>Affects Versions: 1.10.0
>Reporter: Chen Qin
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Native intervaljoin rely on statebackend(e.g rocksdb) to insert/fetch left 
> and right buffer. This design choice reduce minimize heap memory footprint 
> while bounded process throughput of single taskmanager iops to rocksdb access 
> speed. Here at Pinterest, we have some large use cases where developers join 
> large and slow evolving data stream (e.g post updates in last 28 days) with 
> web traffic datastream (e.g post views up to 28 days after given update).
> This post some challenge to current implementation of intervaljoin
>  * partitioned rocksdb needs to keep both updates and views for 28 days, 
> large buffer(especially view stream side) cause rocksdb slow down and lead to 
> overall interval join performance degregate quickly as state build up.
>  * view stream is web scale, even after setting large parallelism it can put 
> lot of pressure on each subtask and backpressure entire job
> In proposed implementation, we plan to introduce two changes
>  * support ProcessJoinFunction settings to opt-in earlier cleanup time of 
> right stream(e.g view stream don't have to stay in buffer for 28 days and 
> wait for update stream to join, related post views happens after update in 
> event time semantic) This optimization can reduce state size to improve 
> rocksdb throughput. If extreme case, user can opt-in in flight join and skip 
> write into right view stream buffer to save iops budget on each subtask
>  * support ProcessJoinFunction settings to expedite keyed lookup of slow 
> changing stream. Instead of every post view pull post updates from rocksdb. 
> user can opt-in and having one side buffer cache available in memory. If a 
> given post update, cache load recent views from right buffer and use 
> sortedMap to find buckets. If a given post view, cache load recent updates 
> from left buffer to memory. When another view for that post arrives, flink 
> save cost of rocksdb access.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16575) develop HBaseCatalog to integrate HBase metadata into Flink

2020-03-13 Thread Bowen Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16575?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-16575:
-
Description: 
develop HBaseCatalog to integrate HBase metadata into Flink

The ticket includes necessary initial investigation to see if it's possible and 
brings practical value, since hbase/elasticsearch are schemaless.
 
If it is valuable, then partition/function/stats/views probably shouldn't be 
implemented, which would be very similar to PostgresCatalog 
([https://github.com/apache/flink/pull/11336]). HiveCatalog can also be a good 
reference.

  was:develop HBaseCatalog to integrate HBase metadata into Flink


> develop HBaseCatalog to integrate HBase metadata into Flink
> ---
>
> Key: FLINK-16575
> URL: https://issues.apache.org/jira/browse/FLINK-16575
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / HBase
>Reporter: Bowen Li
>Priority: Major
>
> develop HBaseCatalog to integrate HBase metadata into Flink
> The ticket includes necessary initial investigation to see if it's possible 
> and brings practical value, since hbase/elasticsearch are schemaless.
>  
> If it is valuable, then partition/function/stats/views probably shouldn't be 
> implemented, which would be very similar to PostgresCatalog 
> ([https://github.com/apache/flink/pull/11336]). HiveCatalog can also be a 
> good reference.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16575) develop HBaseCatalog to integrate HBase metadata into Flink

2020-03-13 Thread Bowen Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058974#comment-17058974
 ] 

Bowen Li commented on FLINK-16575:
--

this ticket would involve initial research to see if it is feasible and has 
practical value

> develop HBaseCatalog to integrate HBase metadata into Flink
> ---
>
> Key: FLINK-16575
> URL: https://issues.apache.org/jira/browse/FLINK-16575
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / HBase
>Reporter: Bowen Li
>Priority: Major
>
> develop HBaseCatalog to integrate HBase metadata into Flink



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392371252
 
 

 ##
 File path: tools/travis/common-logging.sh
 ##
 @@ -0,0 +1,123 @@
+#!/usr/bin/env bash
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  "License"); you may not use this file except in compliance
+#  with the License.  You may obtain a copy of the License at
+#
+#  http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+# Maximum times to retry uploading artifacts file to transfer.sh
+TRANSFER_UPLOAD_MAX_RETRIES=2
+
+# The delay between two retries to upload artifacts file to transfer.sh. The 
default exponential
+# backoff algorithm should be too long for the last several retries.
+TRANSFER_UPLOAD_RETRY_DELAY=5
+
+# E.g. travis-artifacts/apache/flink/1595/1595.1
+UPLOAD_TARGET_PATH="travis-artifacts/${TRAVIS_REPO_SLUG}/${TRAVIS_BUILD_NUMBER}/"
+# These variables are stored as secure variables in '.travis.yml', which are 
generated per repo via
+# the travis command line tool.
+UPLOAD_BUCKET=$ARTIFACTS_AWS_BUCKET
+UPLOAD_ACCESS_KEY=$ARTIFACTS_AWS_ACCESS_KEY
+UPLOAD_SECRET_KEY=$ARTIFACTS_AWS_SECRET_KEY
+
+
+SCRIPT_DIR="`dirname \"$0\"`"
+SCRIPT_DIR="`( cd \"${SCRIPT_DIR}\" && pwd -P)`"
+export FLINK_ROOT="`( cd \"${SCRIPT_DIR}/..\" && pwd -P)`"
 
 Review comment:
   this seems quite brittle


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392367738
 
 

 ##
 File path: tools/azure_controller.sh
 ##
 @@ -53,6 +51,8 @@ print_system_info() {
 
 print_system_info
 
+# enable core dumps
+sudo ulimit -c unlimited
 
 Review comment:
   what does this have to do with core dumps?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392367973
 
 

 ##
 File path: tools/azure_controller.sh
 ##
 @@ -24,8 +24,6 @@ echo $MAVEN_OPTS
 mvn -version
 echo "Commit: $(git rev-parse HEAD)"
 
-
-
 
 Review comment:
   unrelated


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392367035
 
 

 ##
 File path: tools/travis_watchdog.sh
 ##
 @@ -177,6 +177,14 @@ print_stacktraces () {
done
 }
 
+collect_coredumps() {
+   echo "Searching for .dump, .dumpstream and related files in 
$($HERE/../)"
+   for file in `find . -type f -regextype posix-extended -iregex 
'.*\.dump|.*\.dumpstream|.*hs.*\.log|.*/core(.[0-9]+)?$'`; do
 
 Review comment:
   a simpler solution may be to setup the JVM to directly write the dumps into 
`ARTIFACTS_DIR`.
   
https://stackoverflow.com/questions/1880166/is-it-possible-to-specify-where-jvms-crash-dumps-go


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392364911
 
 

 ##
 File path: flink-end-to-end-tests/test-scripts/test-runner-common.sh
 ##
 @@ -46,7 +46,7 @@ function run_test {
 
 function test_error() {
   echo "[FAIL] Test script contains errors."
-  post_test_validation 1 "$description" "$skip_check_exceptions"
+  post_test_validation 1 "$description"
 
 Review comment:
   My impressions is that this will actually obfuscate errors. People pretty 
much only look at the last ~20 lines of the CI output, and don't properly 
investigate the case, resulting in several superfluous JIRAs.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement 
new StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598702521
 
 
   
   ## CI report:
   
   * c1facefebee52a1ac84f188f46b114deae0aad89 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153207095) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392338175
 
 

 ##
 File path: tools/travis/common-logging.sh
 ##
 @@ -0,0 +1,123 @@
+#!/usr/bin/env bash
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  "License"); you may not use this file except in compliance
+#  with the License.  You may obtain a copy of the License at
+#
+#  http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+# Maximum times to retry uploading artifacts file to transfer.sh
+TRANSFER_UPLOAD_MAX_RETRIES=2
+
+# The delay between two retries to upload artifacts file to transfer.sh. The 
default exponential
+# backoff algorithm should be too long for the last several retries.
+TRANSFER_UPLOAD_RETRY_DELAY=5
+
+# E.g. travis-artifacts/apache/flink/1595/1595.1
+UPLOAD_TARGET_PATH="travis-artifacts/${TRAVIS_REPO_SLUG}/${TRAVIS_BUILD_NUMBER}/"
+# These variables are stored as secure variables in '.travis.yml', which are 
generated per repo via
+# the travis command line tool.
+UPLOAD_BUCKET=$ARTIFACTS_AWS_BUCKET
+UPLOAD_ACCESS_KEY=$ARTIFACTS_AWS_ACCESS_KEY
+UPLOAD_SECRET_KEY=$ARTIFACTS_AWS_SECRET_KEY
+
+
+SCRIPT_DIR="`dirname \"$0\"`"
+SCRIPT_DIR="`( cd \"${SCRIPT_DIR}\" && pwd -P)`"
+export FLINK_ROOT="`( cd \"${SCRIPT_DIR}/..\" && pwd -P)`"
+if [ -z "${FLINK_ROOT}" ] ; then
+   # error; for some reason, the path is not accessible
+   # to the script (e.g. permissions re-evaled after suid)
+   exit 1  # fail
+fi
+
+prepare_artifacts() {
+   export ARTIFACTS_DIR="${SCRIPT_DIR}/artifacts"
+
+   mkdir -p $ARTIFACTS_DIR || { echo "FAILURE: cannot create log directory 
'${ARTIFACTS_DIR}'." ; exit 1; }
+}
+
+upload_artifacts() {
+   ARTIFACTS_FILE=${TRAVIS_JOB_NUMBER}.tar.gz
+   if [ ! -z "$TF_BUILD" ] ; then
+   # set proper artifacts file name on Azure Pipelines
+   ARTIFACTS_FILE=${BUILD_BUILDNUMBER}.tar.gz
+   if [ ! -z "$MODULE" ] ; then
+   ARTIFACTS_FILE=${BUILD_BUILDNUMBER}-$(echo $MODULE | tr 
-dc '[:alnum:]\n\r').tar.gz
+   fi
+   fi
+
+   echo "PRODUCED build artifacts."
 
 Review comment:
   as usual this would be easier to review if the moving of code were put into 
a separate commit.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11372: [FLINK-16480] Improve error reporting on AZP

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11372: [FLINK-16480] Improve 
error reporting on AZP
URL: https://github.com/apache/flink/pull/11372#discussion_r392337003
 
 

 ##
 File path: flink-end-to-end-tests/test-scripts/test-runner-common.sh
 ##
 @@ -46,7 +46,7 @@ function run_test {
 
 function test_error() {
   echo "[FAIL] Test script contains errors."
-  post_test_validation 1 "$description" "$skip_check_exceptions"
+  post_test_validation 1 "$description"
 
 Review comment:
   do you know whether the bash tests upload the logs for failed tests?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-14121) Upgrade commons-compress to 1.20 due to CVE

2020-03-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao closed FLINK-14121.

Resolution: Fixed

master: 0477368813fba33f0f9aa5ff82304516c499ed41

> Upgrade commons-compress to 1.20 due to CVE
> ---
>
> Key: FLINK-14121
> URL: https://issues.apache.org/jira/browse/FLINK-14121
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream, Build System, Release System
>Affects Versions: 1.9.0
>Reporter: John Lonergan
>Assignee: Niels Basjes
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> See 
> https://commons.apache.org/proper/commons-compress/security-reports.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14121) Upgrade commons-compress to 1.20 due to CVE

2020-03-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-14121:
-
Summary: Upgrade commons-compress to 1.20 due to CVE  (was: Upgrade 
commons-compress to 1.19 due to CVE)

> Upgrade commons-compress to 1.20 due to CVE
> ---
>
> Key: FLINK-14121
> URL: https://issues.apache.org/jira/browse/FLINK-14121
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream, Build System, Release System
>Affects Versions: 1.9.0
>Reporter: John Lonergan
>Assignee: Niels Basjes
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> See 
> https://commons.apache.org/proper/commons-compress/security-reports.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type inference for temporal table functions

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type 
inference for temporal table functions
URL: https://github.com/apache/flink/pull/11253#issuecomment-592517598
 
 
   
   ## CI report:
   
   * 4e1c4f61f4e477262eb3d1d4fa67c97bcec1b85c Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153201488) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-14121) Upgrade commons-compress to 1.19 due to CVE

2020-03-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-14121:
-
Fix Version/s: 1.11.0

> Upgrade commons-compress to 1.19 due to CVE
> ---
>
> Key: FLINK-14121
> URL: https://issues.apache.org/jira/browse/FLINK-14121
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream, Build System, Release System
>Affects Versions: 1.9.0
>Reporter: John Lonergan
>Assignee: Niels Basjes
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> See 
> https://commons.apache.org/proper/commons-compress/security-reports.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] GJL closed pull request #11333: [FLINK-14121] Update commons-compress because of CVE-2019-12402

2020-03-13 Thread GitBox
GJL closed pull request #11333: [FLINK-14121] Update commons-compress because 
of CVE-2019-12402
URL: https://github.com/apache/flink/pull/11333
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (FLINK-16553) Missing KafkaFetcher topic/partition metrics

2020-03-13 Thread Tarush Grover (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058893#comment-17058893
 ] 

Tarush Grover edited comment on FLINK-16553 at 3/13/20, 4:24 PM:
-

[~aljoscha] I will come up with implementation document and share it with all.


was (Author: app-tarush):
[~aljoscha] I will come up with implementation document and share with all.

> Missing KafkaFetcher topic/partition metrics
> 
>
> Key: FLINK-16553
> URL: https://issues.apache.org/jira/browse/FLINK-16553
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka, Runtime / Metrics
>Reporter: Fabian Paul
>Assignee: Tarush Grover
>Priority: Major
>
> When using the Kafka universal connector, currently not all KafkaFetcher 
> metrics 
> ([link|https://github.com/apache/kafka/blob/trunk/clients/src/main/java/org/apache/kafka/clients/consumer/internals/FetcherMetricsRegistry.java])
>  which are exposed through the KafkaConsumer are accessible within the Flink 
> metrics system.
> Especially, all metrics which are related to topics and partitions are not 
> available. The KafkaConsumer internally only registers those metrics after it 
> has fetched some records.
> Unfortunately, at the moment Flink only checks the available metrics right 
> after the initialization of the KafkaConsumer when no records are polled, yet.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16553) Missing KafkaFetcher topic/partition metrics

2020-03-13 Thread Tarush Grover (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058893#comment-17058893
 ] 

Tarush Grover commented on FLINK-16553:
---

[~aljoscha] I will come up with implementation document and share with all.

> Missing KafkaFetcher topic/partition metrics
> 
>
> Key: FLINK-16553
> URL: https://issues.apache.org/jira/browse/FLINK-16553
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka, Runtime / Metrics
>Reporter: Fabian Paul
>Assignee: Tarush Grover
>Priority: Major
>
> When using the Kafka universal connector, currently not all KafkaFetcher 
> metrics 
> ([link|https://github.com/apache/kafka/blob/trunk/clients/src/main/java/org/apache/kafka/clients/consumer/internals/FetcherMetricsRegistry.java])
>  which are exposed through the KafkaConsumer are accessible within the Flink 
> metrics system.
> Especially, all metrics which are related to topics and partitions are not 
> available. The KafkaConsumer internally only registers those metrics after it 
> has fetched some records.
> Unfortunately, at the moment Flink only checks the available metrics right 
> after the initialization of the KafkaConsumer when no records are polled, yet.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11404: [FLINK-13000][tests] Remove unused JobID parameters

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11404: [FLINK-13000][tests] Remove unused 
JobID parameters
URL: https://github.com/apache/flink/pull/11404#issuecomment-598746532
 
 
   
   ## CI report:
   
   * a7f81066eb1d939e9d6c492895bd5e0eb021360b Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153198834) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16590) flink-oss-fs-hadoop: Not all dependencies in NOTICE file are bundled

2020-03-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-16590:
-
Description: NOTICE file in flink-oss-fs-hadoop lists 
{{org.apache.commons:commons-compress}} as a bundled dependency which is not 
correct. There are likely other dependencies that are wrongly listed in the 
NOTICE file.  (was: NOTICE file in flink-oss-fs-hadoop lists
{{code}}
org.apache.commons:commons-compress
{{code}}

as a bundled dependency which is not correct. There are likely other 
dependencies that are wrongly listed in the NOTICE file.)

> flink-oss-fs-hadoop: Not all dependencies in NOTICE file are bundled
> 
>
> Key: FLINK-16590
> URL: https://issues.apache.org/jira/browse/FLINK-16590
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem, Release System
>Affects Versions: 1.11.0
>Reporter: Gary Yao
>Priority: Critical
> Fix For: 1.11.0
>
>
> NOTICE file in flink-oss-fs-hadoop lists 
> {{org.apache.commons:commons-compress}} as a bundled dependency which is not 
> correct. There are likely other dependencies that are wrongly listed in the 
> NOTICE file.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16590) flink-oss-fs-hadoop: Not all dependencies in NOTICE file are bundled

2020-03-13 Thread Gary Yao (Jira)
Gary Yao created FLINK-16590:


 Summary: flink-oss-fs-hadoop: Not all dependencies in NOTICE file 
are bundled
 Key: FLINK-16590
 URL: https://issues.apache.org/jira/browse/FLINK-16590
 Project: Flink
  Issue Type: Bug
  Components: Connectors / FileSystem, Release System
Affects Versions: 1.11.0
Reporter: Gary Yao
 Fix For: 1.11.0


NOTICE file in flink-oss-fs-hadoop lists
{{code}}
org.apache.commons:commons-compress
{{code}}

as a bundled dependency which is not correct. There are likely other 
dependencies that are wrongly listed in the NOTICE file.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16408) Bind user code class loader to lifetime of a slot

2020-03-13 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16408?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058862#comment-17058862
 ] 

Stephan Ewen commented on FLINK-16408:
--

I think it should be okay to fix in a 1.10.x release, but want to double check. 
It may have some implications, for example statics not being re-initialized in 
restart, because the same classloader is being used.

I would assume this to be okay, in fact, really helpful in most cases, but 
wanted to check if someone would see a concern with changing this behavior.

> Bind user code class loader to lifetime of a slot
> -
>
> Key: FLINK-16408
> URL: https://issues.apache.org/jira/browse/FLINK-16408
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.2, 1.10.0
>Reporter: Till Rohrmann
>Priority: Critical
> Fix For: 1.10.1, 1.11.0
>
>
> In order to avoid class leaks due to creating multiple user code class 
> loaders and loading class multiple times in a recovery case, I would suggest 
> to bind the lifetime of a user code class loader to the lifetime of a slot. 
> More precisely, the user code class loader should live at most as long as the 
> slot which is using it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16589) Flink Table SQL fails/crashes with big queries with lots of fields

2020-03-13 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-16589:

Priority: Critical  (was: Major)

> Flink Table SQL fails/crashes with big queries with lots of fields
> --
>
> Key: FLINK-16589
> URL: https://issues.apache.org/jira/browse/FLINK-16589
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Viet Pham
>Assignee: Benchao Li
>Priority: Critical
> Fix For: 1.10.1, 1.11.0
>
>
> Hi,
> My use case is a streaming application with a few streaming tables.
> I was trying to build a SELECT query (and registering it as a temporary view) 
> with about 200 fields/expressions out of another streaming table. The 
> application is successfully submitted to Flink cluster. However the worker 
> processes keep crashing, with the exception as quoted below. 
> It clearly mentioned in the log that this is a bug, so I fire this ticket. By 
> the way, if I lower the number of fields down to 100 then it works nicely.
> Please advice.
> Thanks a lot for all the efforts bring Flink up. It is really amazing!
> {code:java}
> java.lang.RuntimeException: Could not instantiate generated class 
> 'GroupAggsHandler$9687'at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:57)
> at 
> org.apache.flink.table.runtime.operators.aggregate.MiniBatchGroupAggFunction.open(MiniBatchGroupAggFunction.java:136)
> at 
> org.apache.flink.table.runtime.operators.bundle.AbstractMapBundleOperator.open(AbstractMapBundleOperator.java:84)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeStateAndOpen(StreamTask.java:1007)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:454)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:94)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:449)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:461)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:68)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:78)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:52)
> ... 10 moreCaused by: 
> org.apache.flink.shaded.guava18.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2203)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache.get(LocalCache.java:3937)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4739)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:66)
> ... 12 moreCaused by: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:81)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:66)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4742)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
> ... 15 moreCaused by: org.codehaus.janino.InternalCompilerException: 
> Compiling "GroupAggsHandler$9687": Code of method 
> "retract(Lorg/apache/flink/table/dataformat/BaseRow;)V" of class 
> "GroupAggsHandler$9687" grows beyond 64 KBat 
> org.codehaus.janino

[jira] [Updated] (FLINK-16589) Flink Table SQL fails/crashes with big queries with lots of fields

2020-03-13 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-16589:

Fix Version/s: 1.11.0
   1.10.1

> Flink Table SQL fails/crashes with big queries with lots of fields
> --
>
> Key: FLINK-16589
> URL: https://issues.apache.org/jira/browse/FLINK-16589
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Viet Pham
>Assignee: Benchao Li
>Priority: Major
> Fix For: 1.10.1, 1.11.0
>
>
> Hi,
> My use case is a streaming application with a few streaming tables.
> I was trying to build a SELECT query (and registering it as a temporary view) 
> with about 200 fields/expressions out of another streaming table. The 
> application is successfully submitted to Flink cluster. However the worker 
> processes keep crashing, with the exception as quoted below. 
> It clearly mentioned in the log that this is a bug, so I fire this ticket. By 
> the way, if I lower the number of fields down to 100 then it works nicely.
> Please advice.
> Thanks a lot for all the efforts bring Flink up. It is really amazing!
> {code:java}
> java.lang.RuntimeException: Could not instantiate generated class 
> 'GroupAggsHandler$9687'at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:57)
> at 
> org.apache.flink.table.runtime.operators.aggregate.MiniBatchGroupAggFunction.open(MiniBatchGroupAggFunction.java:136)
> at 
> org.apache.flink.table.runtime.operators.bundle.AbstractMapBundleOperator.open(AbstractMapBundleOperator.java:84)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeStateAndOpen(StreamTask.java:1007)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:454)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:94)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:449)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:461)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:68)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:78)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:52)
> ... 10 moreCaused by: 
> org.apache.flink.shaded.guava18.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2203)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache.get(LocalCache.java:3937)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4739)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:66)
> ... 12 moreCaused by: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:81)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:66)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4742)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
> ... 15 moreCaused by: org.codehaus.janino.InternalCompilerException: 
> Compiling "GroupAggsHandler$9687": Code of method 
> "retract(Lorg/apache/flink/table/dataformat/BaseRow;)V" of class 
> "GroupAggsHandler$9687" grows beyond 64 KBat 
> org.cod

[jira] [Assigned] (FLINK-16589) Flink Table SQL fails/crashes with big queries with lots of fields

2020-03-13 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-16589:
---

Assignee: Benchao Li

> Flink Table SQL fails/crashes with big queries with lots of fields
> --
>
> Key: FLINK-16589
> URL: https://issues.apache.org/jira/browse/FLINK-16589
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Viet Pham
>Assignee: Benchao Li
>Priority: Major
>
> Hi,
> My use case is a streaming application with a few streaming tables.
> I was trying to build a SELECT query (and registering it as a temporary view) 
> with about 200 fields/expressions out of another streaming table. The 
> application is successfully submitted to Flink cluster. However the worker 
> processes keep crashing, with the exception as quoted below. 
> It clearly mentioned in the log that this is a bug, so I fire this ticket. By 
> the way, if I lower the number of fields down to 100 then it works nicely.
> Please advice.
> Thanks a lot for all the efforts bring Flink up. It is really amazing!
> {code:java}
> java.lang.RuntimeException: Could not instantiate generated class 
> 'GroupAggsHandler$9687'at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:57)
> at 
> org.apache.flink.table.runtime.operators.aggregate.MiniBatchGroupAggFunction.open(MiniBatchGroupAggFunction.java:136)
> at 
> org.apache.flink.table.runtime.operators.bundle.AbstractMapBundleOperator.open(AbstractMapBundleOperator.java:84)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeStateAndOpen(StreamTask.java:1007)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:454)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:94)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:449)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:461)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:68)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:78)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:52)
> ... 10 moreCaused by: 
> org.apache.flink.shaded.guava18.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2203)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache.get(LocalCache.java:3937)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4739)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:66)
> ... 12 moreCaused by: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:81)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:66)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4742)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
> ... 15 moreCaused by: org.codehaus.janino.InternalCompilerException: 
> Compiling "GroupAggsHandler$9687": Code of method 
> "retract(Lorg/apache/flink/table/dataformat/BaseRow;)V" of class 
> "GroupAggsHandler$9687" grows beyond 64 KBat 
> org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:382)

[jira] [Commented] (FLINK-16589) Flink Table SQL fails/crashes with big queries with lots of fields

2020-03-13 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058859#comment-17058859
 ] 

Jark Wu commented on FLINK-16589:
-

Yes. We didn't split code for GroupAggsHandler, but that can be done as 
[~libenchao] said, even though it is a complete solution... 

Assigned this issue to you [~libenchao].

> Flink Table SQL fails/crashes with big queries with lots of fields
> --
>
> Key: FLINK-16589
> URL: https://issues.apache.org/jira/browse/FLINK-16589
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Viet Pham
>Assignee: Benchao Li
>Priority: Major
> Fix For: 1.10.1, 1.11.0
>
>
> Hi,
> My use case is a streaming application with a few streaming tables.
> I was trying to build a SELECT query (and registering it as a temporary view) 
> with about 200 fields/expressions out of another streaming table. The 
> application is successfully submitted to Flink cluster. However the worker 
> processes keep crashing, with the exception as quoted below. 
> It clearly mentioned in the log that this is a bug, so I fire this ticket. By 
> the way, if I lower the number of fields down to 100 then it works nicely.
> Please advice.
> Thanks a lot for all the efforts bring Flink up. It is really amazing!
> {code:java}
> java.lang.RuntimeException: Could not instantiate generated class 
> 'GroupAggsHandler$9687'at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:57)
> at 
> org.apache.flink.table.runtime.operators.aggregate.MiniBatchGroupAggFunction.open(MiniBatchGroupAggFunction.java:136)
> at 
> org.apache.flink.table.runtime.operators.bundle.AbstractMapBundleOperator.open(AbstractMapBundleOperator.java:84)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeStateAndOpen(StreamTask.java:1007)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:454)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:94)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:449)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:461)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:68)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:78)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:52)
> ... 10 moreCaused by: 
> org.apache.flink.shaded.guava18.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2203)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache.get(LocalCache.java:3937)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4739)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:66)
> ... 12 moreCaused by: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:81)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:66)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4742)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
> ... 15 moreCaused by: org.codehaus.janino.InternalCompilerException: 

[GitHub] [flink] AHeise commented on issue #11303: [FLINK-16245] Decoupling user classloader from context classloader.

2020-03-13 Thread GitBox
AHeise commented on issue #11303: [FLINK-16245] Decoupling user classloader 
from context classloader.
URL: https://github.com/apache/flink/pull/11303#issuecomment-598786153
 
 
   @flinkbot run azure


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Assigned] (FLINK-16047) Blink planner produces wrong aggregate results with state clean up

2020-03-13 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-16047:
---

Assignee: Jark Wu

> Blink planner produces wrong aggregate results with state clean up
> --
>
> Key: FLINK-16047
> URL: https://issues.apache.org/jira/browse/FLINK-16047
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: Timo Walther
>Assignee: Jark Wu
>Priority: Critical
> Fix For: 1.9.3, 1.10.1, 1.11.0
>
>
> It seems that FLINK-10674 has not been ported to the Blink planner.
> Because state clean up happens in processing time, it might be the case that 
> retractions are arriving after the state has been cleaned up. Before these 
> changes, a new accumulator was created and invalid retraction messages were 
> emitted. This change drops retraction messages for which no accumulator 
> exists.
> These lines are missing in 
> {{org.apache.flink.table.runtime.operators.aggregate.GroupAggFunction}}:
> {code}
> if (null == accumulators) {
>   // Don't create a new accumulator for a retraction message. This
>   // might happen if the retraction message is the first message for the
>   // key or after a state clean up.
>   if (!inputC.change) {
> return
>   }
>   // first accumulate message
>   firstRow = true
>   accumulators = function.createAccumulators()
> } else {
>   firstRow = false
> }
> {code}
> The bug has not been verified. I spotted it only by looking at the code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16047) Blink planner produces wrong aggregate results with state clean up

2020-03-13 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16047?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058858#comment-17058858
 ] 

Jark Wu commented on FLINK-16047:
-

I will pick up this issue and submit a PR soon.

> Blink planner produces wrong aggregate results with state clean up
> --
>
> Key: FLINK-16047
> URL: https://issues.apache.org/jira/browse/FLINK-16047
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: Timo Walther
>Priority: Critical
> Fix For: 1.9.3, 1.10.1, 1.11.0
>
>
> It seems that FLINK-10674 has not been ported to the Blink planner.
> Because state clean up happens in processing time, it might be the case that 
> retractions are arriving after the state has been cleaned up. Before these 
> changes, a new accumulator was created and invalid retraction messages were 
> emitted. This change drops retraction messages for which no accumulator 
> exists.
> These lines are missing in 
> {{org.apache.flink.table.runtime.operators.aggregate.GroupAggFunction}}:
> {code}
> if (null == accumulators) {
>   // Don't create a new accumulator for a retraction message. This
>   // might happen if the retraction message is the first message for the
>   // key or after a state clean up.
>   if (!inputC.change) {
> return
>   }
>   // first accumulate message
>   firstRow = true
>   accumulators = function.createAccumulators()
> } else {
>   firstRow = false
> }
> {code}
> The bug has not been verified. I spotted it only by looking at the code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in respective constructors

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build 
ResultSubpartitionInfo and InputChannelInfo in respective constructors
URL: https://github.com/apache/flink/pull/11400#issuecomment-598611898
 
 
   
   ## CI report:
   
   * a86f1e2e6785237893201dac302bc50041610d73 UNKNOWN
   * 5216b6412ed32669643e576b1879af2122f43e4b Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153193702) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6290)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement 
new StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598702521
 
 
   
   ## CI report:
   
   * c7595ee5eb7ac60dae52bcd1368878d70d372533 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153184710) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6287)
 
   * c1facefebee52a1ac84f188f46b114deae0aad89 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153207095) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement 
new StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598702521
 
 
   
   ## CI report:
   
   * c7595ee5eb7ac60dae52bcd1368878d70d372533 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153184710) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6287)
 
   * c1facefebee52a1ac84f188f46b114deae0aad89 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14348) YarnFileStageTestS3ITCase. testRecursiveUploadForYarnS3a fails to delete files

2020-03-13 Thread Aljoscha Krettek (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14348?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058832#comment-17058832
 ] 

Aljoscha Krettek commented on FLINK-14348:
--

Same test, but different failure: 
https://dev.azure.com/aljoschakrettek/Flink/_build/results?buildId=59&view=logs&j=764762df-f65b-572b-3d5c-65518c777be4&t=da3c2718-4b76-56bf-ef25-cd33ea381f78

> YarnFileStageTestS3ITCase. testRecursiveUploadForYarnS3a fails to delete files
> --
>
> Key: FLINK-14348
> URL: https://issues.apache.org/jira/browse/FLINK-14348
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.9.0
>Reporter: Caizhi Weng
>Priority: Major
>
> YarnFileStageTestS3ITCase. testRecursiveUploadForYarnS3a fails with the 
> following exceptions:
> {code:java}
> 15:25:07.359 [ERROR] 
> testRecursiveUploadForYarnS3a(org.apache.flink.yarn.YarnFileStageTestS3ITCase)
>   Time elapsed: 10.808 s  <<< 
> ERROR!24649org.apache.hadoop.fs.s3a.AWSS3IOException: delete on 
> s3a://[secure]/temp/tests-3565b11f-e9be-4213-a98d-0f0ecd123783/testYarn-s3a: 
> com.amazonaws.services.s3.model.MultiObjectDeleteException: One or more 
> objects could not be deleted (Service: null; Status Code: 200; Error Code: 
> null; Request ID: 2D1AE3D999528C34; S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=),
>  S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=: 
> One or more objects could not be deleted (Service: null; Status Code: 200; 
> Error Code: null; Request ID: 2D1AE3D999528C34; S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=)24650
>at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:159)24651
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3a(YarnFileStageTestS3ITCase.java:190)24652Caused
>  by: com.amazonaws.services.s3.model.MultiObjectDeleteException: One or more 
> objects could not be deleted (Service: null; Status Code: 200; Error Code: 
> null; Request ID: 2D1AE3D999528C34; S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=)24653
>at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:159)24654
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3a(YarnFileStageTestS3ITCase.java:190){code}
> Travis log: [https://travis-ci.org/apache/flink/jobs/595082651]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14348) YarnFileStageTestS3ITCase. testRecursiveUploadForYarnS3a fails to delete files

2020-03-13 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek updated FLINK-14348:
-
Component/s: Build System / Azure Pipelines

> YarnFileStageTestS3ITCase. testRecursiveUploadForYarnS3a fails to delete files
> --
>
> Key: FLINK-14348
> URL: https://issues.apache.org/jira/browse/FLINK-14348
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / Azure Pipelines, Tests
>Affects Versions: 1.9.0
>Reporter: Caizhi Weng
>Priority: Major
>
> YarnFileStageTestS3ITCase. testRecursiveUploadForYarnS3a fails with the 
> following exceptions:
> {code:java}
> 15:25:07.359 [ERROR] 
> testRecursiveUploadForYarnS3a(org.apache.flink.yarn.YarnFileStageTestS3ITCase)
>   Time elapsed: 10.808 s  <<< 
> ERROR!24649org.apache.hadoop.fs.s3a.AWSS3IOException: delete on 
> s3a://[secure]/temp/tests-3565b11f-e9be-4213-a98d-0f0ecd123783/testYarn-s3a: 
> com.amazonaws.services.s3.model.MultiObjectDeleteException: One or more 
> objects could not be deleted (Service: null; Status Code: 200; Error Code: 
> null; Request ID: 2D1AE3D999528C34; S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=),
>  S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=: 
> One or more objects could not be deleted (Service: null; Status Code: 200; 
> Error Code: null; Request ID: 2D1AE3D999528C34; S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=)24650
>at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:159)24651
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3a(YarnFileStageTestS3ITCase.java:190)24652Caused
>  by: com.amazonaws.services.s3.model.MultiObjectDeleteException: One or more 
> objects could not be deleted (Service: null; Status Code: 200; Error Code: 
> null; Request ID: 2D1AE3D999528C34; S3 Extended Request ID: 
> zIX1QsAcsY1ZYSDOeCaYsGJ4bz0NJTy2kw0EYmlJr8Kb7pM8OPmhAKO5XHI26xiOi2tIkTIoBwg=)24653
>at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarn(YarnFileStageTestS3ITCase.java:159)24654
>   at 
> org.apache.flink.yarn.YarnFileStageTestS3ITCase.testRecursiveUploadForYarnS3a(YarnFileStageTestS3ITCase.java:190){code}
> Travis log: [https://travis-ci.org/apache/flink/jobs/595082651]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-11088) Allow pre-install Kerberos authentication keytab discovery on YARN

2020-03-13 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-11088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek closed FLINK-11088.

Resolution: Implemented

master: 0920e6447283579551d14a0fed726f8b8d73bd3f

> Allow pre-install Kerberos authentication keytab discovery on YARN
> --
>
> Key: FLINK-11088
> URL: https://issues.apache.org/jira/browse/FLINK-11088
> Project: Flink
>  Issue Type: Sub-task
>  Components: Deployment / YARN
>Reporter: Rong Rong
>Assignee: Rong Rong
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Currently flink-yarn assumes keytab is shipped as application master 
> environment local resource on client side and will be distributed to all the 
> TMs. This does not work for YARN proxy user mode [1] since proxy user or 
> super user might not have access to actual users' keytab, but can request 
> delegation tokens on users' behalf. 
> Based on the type of security options for long-living YARN service[2], we 
> propose to have the keytab file path discovery configurable depending on the 
> launch mode of the YARN client. 
> Reference: 
> [1] 
> https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/Superusers.html
> [2] 
> https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YarnApplicationSecurity.html#Securing_Long-lived_YARN_Services



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] aljoscha commented on issue #7702: [FLINK-11088][Security][YARN] Allow YARN to discover pre-installed keytab files

2020-03-13 Thread GitBox
aljoscha commented on issue #7702: [FLINK-11088][Security][YARN] Allow YARN to 
discover pre-installed keytab files
URL: https://github.com/apache/flink/pull/7702#issuecomment-598772667
 
 
   I merged this. Thanks a lot for the good collaboration on this. 😃


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] aljoscha closed pull request #7702: [FLINK-11088][Security][YARN] Allow YARN to discover pre-installed keytab files

2020-03-13 Thread GitBox
aljoscha closed pull request #7702: [FLINK-11088][Security][YARN] Allow YARN to 
discover pre-installed keytab files
URL: https://github.com/apache/flink/pull/7702
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] rkhachatryan commented on a change in pull request #11400: [FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in respective constructors

2020-03-13 Thread GitBox
rkhachatryan commented on a change in pull request #11400: 
[FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in 
respective constructors
URL: https://github.com/apache/flink/pull/11400#discussion_r392291325
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/NettyShuffleEnvironment.java
 ##
 @@ -217,6 +218,7 @@ public ShuffleIOOwnerContext createShuffleIOOwnerContext(
for (InputGateDeploymentDescriptor igdd : 
inputGateDeploymentDescriptors) {
SingleInputGate inputGate = 
singleInputGateFactory.create(
ownerContext.getOwnerName(),
+   counter,
 
 Review comment:
   Not sure if I got you right.
   
   I meant to change the type from `Collection` to `List` for:
   1. `inputGateDeploymentDescriptors` parameter in 
`NettyShuffleEnvironment#createInputGates`
   1. `TaskDeploymentDescriptor#inputGates` field
   1. `inputGateDeploymentDescriptors` parameter in `Task` constructor
   1. `inputGateDeploymentDescriptors` parameter in `TaskDeploymentDescriptor` 
constructor
   
   Without it, it's easy to break the ordering, at least for me :)
   
   I think it should be the same PR because requirement for ordering was added 
here.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type inference for temporal table functions

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type 
inference for temporal table functions
URL: https://github.com/apache/flink/pull/11253#issuecomment-592517598
 
 
   
   ## CI report:
   
   * d1368ad6d7262321e92a1e8c70211bc16a61dd05 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153170857) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6284)
 
   * 4e1c4f61f4e477262eb3d1d4fa67c97bcec1b85c Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153201488) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11402: [FLINK-16166][build] Append additional javadoc options

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11402: [FLINK-16166][build] Append 
additional javadoc options
URL: https://github.com/apache/flink/pull/11402#issuecomment-598690738
 
 
   
   ## CI report:
   
   * 3f91303e7b54b0790f556d22da4fedc4872ac137 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153178906) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6286)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11398: [FLINK-16413]Reduce hive source parallelism when limit push down

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11398: [FLINK-16413]Reduce hive source 
parallelism when limit push down
URL: https://github.com/apache/flink/pull/11398#issuecomment-598579739
 
 
   
   ## CI report:
   
   * 726c57d3f58bcf3adba172273a8b207e5f0281fb Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153201637) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11404: [FLINK-13000][tests] Remove unused JobID parameters

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11404: [FLINK-13000][tests] Remove unused 
JobID parameters
URL: https://github.com/apache/flink/pull/11404#issuecomment-598746532
 
 
   
   ## CI report:
   
   * a7f81066eb1d939e9d6c492895bd5e0eb021360b Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153198834) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11398: [FLINK-16413]Reduce hive source parallelism when limit push down

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11398: [FLINK-16413]Reduce hive source 
parallelism when limit push down
URL: https://github.com/apache/flink/pull/11398#issuecomment-598579739
 
 
   
   ## CI report:
   
   * 4fe5672c65d62cb5e1854db4bf3404d1f4a5fdec Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153129571) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6272)
 
   * 726c57d3f58bcf3adba172273a8b207e5f0281fb UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in respective constructors

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build 
ResultSubpartitionInfo and InputChannelInfo in respective constructors
URL: https://github.com/apache/flink/pull/11400#issuecomment-598611898
 
 
   
   ## CI report:
   
   * a86f1e2e6785237893201dac302bc50041610d73 UNKNOWN
   * 5216b6412ed32669643e576b1879af2122f43e4b Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153193702) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6290)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type inference for temporal table functions

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type 
inference for temporal table functions
URL: https://github.com/apache/flink/pull/11253#issuecomment-592517598
 
 
   
   ## CI report:
   
   * d1368ad6d7262321e92a1e8c70211bc16a61dd05 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153170857) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6284)
 
   * 4e1c4f61f4e477262eb3d1d4fa67c97bcec1b85c UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] twalthr commented on issue #11253: [FLINK-16336][table] Add new type inference for temporal table functions

2020-03-13 Thread GitBox
twalthr commented on issue #11253: [FLINK-16336][table] Add new type inference 
for temporal table functions
URL: https://github.com/apache/flink/pull/11253#issuecomment-598753216
 
 
   @flinkbot run travis
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] aljoscha commented on issue #11253: [FLINK-16336][table] Add new type inference for temporal table functions

2020-03-13 Thread GitBox
aljoscha commented on issue #11253: [FLINK-16336][table] Add new type inference 
for temporal table functions
URL: https://github.com/apache/flink/pull/11253#issuecomment-598750105
 
 
   Lools good to merge!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-10742) Let Netty use Flink's buffers directly in credit-based mode

2020-03-13 Thread Zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-10742?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijiang updated FLINK-10742:
-
Release Note: After FLINK-10742, while reading data stream from network on 
receiver side, it would reuse the Flink's buffers directly to avoid extra copy 
from Netty's ByteBuf to Flink's buffer. Therefore the Netty internal memory 
overhead is greatly reduced, which can solve the potential direct 
OutOfMemoryError specially in large-scale jobs.

> Let Netty use Flink's buffers directly in credit-based mode
> ---
>
> Key: FLINK-10742
> URL: https://issues.apache.org/jira/browse/FLINK-10742
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Affects Versions: 1.7.0
>Reporter: Nico Kruber
>Assignee: Yun Gao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> For credit-based flow control, we always have buffers available for data that 
> is sent to use. We could thus use them directly and not copy the network 
> stream into Netty buffers first and then into our buffers.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #11404: [FLINK-13000][tests] Remove unused JobID parameters

2020-03-13 Thread GitBox
flinkbot commented on issue #11404: [FLINK-13000][tests] Remove unused JobID 
parameters
URL: https://github.com/apache/flink/pull/11404#issuecomment-598746532
 
 
   
   ## CI report:
   
   * a7f81066eb1d939e9d6c492895bd5e0eb021360b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #7702: [FLINK-11088][Security][YARN] Allow YARN to discover pre-installed keytab files

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #7702: [FLINK-11088][Security][YARN] Allow 
YARN to discover pre-installed keytab files
URL: https://github.com/apache/flink/pull/7702#issuecomment-572195960
 
 
   
   ## CI report:
   
   * d4a17beb028586a4b8b06a2eebfe7aa21ccf4fbb UNKNOWN
   * 0e955bd5b9f490a8ed66013ccd8ce3f512955513 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153193280) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6289)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in respective constructors

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build 
ResultSubpartitionInfo and InputChannelInfo in respective constructors
URL: https://github.com/apache/flink/pull/11400#issuecomment-598611898
 
 
   
   ## CI report:
   
   * a86f1e2e6785237893201dac302bc50041610d73 UNKNOWN
   * 3c660599680944e8c6f2b3b0ceaf6975b03fc5c4 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153164755) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6282)
 
   * 5216b6412ed32669643e576b1879af2122f43e4b Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153193702) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6290)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #11404: [FLINK-13000][tests] Remove unused JobID parameters

2020-03-13 Thread GitBox
flinkbot commented on issue #11404: [FLINK-13000][tests] Remove unused JobID 
parameters
URL: https://github.com/apache/flink/pull/11404#issuecomment-598737637
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit a7f81066eb1d939e9d6c492895bd5e0eb021360b (Fri Mar 13 
14:08:39 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #7702: [FLINK-11088][Security][YARN] Allow YARN to discover pre-installed keytab files

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #7702: [FLINK-11088][Security][YARN] Allow 
YARN to discover pre-installed keytab files
URL: https://github.com/apache/flink/pull/7702#issuecomment-572195960
 
 
   
   ## CI report:
   
   * d4a17beb028586a4b8b06a2eebfe7aa21ccf4fbb UNKNOWN
   * ae58c4fc1dcbe3ea0fa70326c09803c3e5792ede Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153089193) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6253)
 
   * 0e955bd5b9f490a8ed66013ccd8ce3f512955513 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153193280) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6289)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13000) Remove JobID argument from SimpleSlotProvider

2020-03-13 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-13000:
---
Labels: pull-request-available  (was: )

> Remove JobID argument from SimpleSlotProvider
> -
>
> Key: FLINK-13000
> URL: https://issues.apache.org/jira/browse/FLINK-13000
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination, Tests
>Affects Versions: 1.9.0
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> The JobID constructor argument is unused and can be removed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zentol opened a new pull request #11404: [FLINK-13000][tests] Remove unused JobID parameters

2020-03-13 Thread GitBox
zentol opened a new pull request #11404: [FLINK-13000][tests] Remove unused 
JobID parameters
URL: https://github.com/apache/flink/pull/11404
 
 
   Removes unused JobID parameters from the `SimpleSlotProvider` and 
`ExecutionGraphTestUtils#createSimpleTestGraph`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhijiangW commented on a change in pull request #11400: [FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in respective constructors

2020-03-13 Thread GitBox
zhijiangW commented on a change in pull request #11400: [FLINK-16586][network] 
Build ResultSubpartitionInfo and InputChannelInfo in respective constructors
URL: https://github.com/apache/flink/pull/11400#discussion_r392247611
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/NettyShuffleEnvironment.java
 ##
 @@ -217,6 +218,7 @@ public ShuffleIOOwnerContext createShuffleIOOwnerContext(
for (InputGateDeploymentDescriptor igdd : 
inputGateDeploymentDescriptors) {
SingleInputGate inputGate = 
singleInputGateFactory.create(
ownerContext.getOwnerName(),
+   counter,
 
 Review comment:
   I guess your suggestion is for unifying all the related data structures 
(collection, array) from `TaskDeploymentDescriptor.inputGates` to `List`?
   
   Actually I do not see any correctness or performance problems ATM. This 
unification really spans cross multiple components from scheduler to task 
stack, and I do think it is really necessary to refactor it now. Or if we want 
to do that, it should be a separate ticket out of this PR scope. We only make 
use of previous `counter` to give a correct gate index while creating in 
iteration for this PR motivation. WDYT?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16047) Blink planner produces wrong aggregate results with state clean up

2020-03-13 Thread Yu Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16047?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058748#comment-17058748
 ] 

Yu Li commented on FLINK-16047:
---

[~jark] Any update on this one? Thanks.

> Blink planner produces wrong aggregate results with state clean up
> --
>
> Key: FLINK-16047
> URL: https://issues.apache.org/jira/browse/FLINK-16047
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: Timo Walther
>Priority: Critical
> Fix For: 1.9.3, 1.10.1, 1.11.0
>
>
> It seems that FLINK-10674 has not been ported to the Blink planner.
> Because state clean up happens in processing time, it might be the case that 
> retractions are arriving after the state has been cleaned up. Before these 
> changes, a new accumulator was created and invalid retraction messages were 
> emitted. This change drops retraction messages for which no accumulator 
> exists.
> These lines are missing in 
> {{org.apache.flink.table.runtime.operators.aggregate.GroupAggFunction}}:
> {code}
> if (null == accumulators) {
>   // Don't create a new accumulator for a retraction message. This
>   // might happen if the retraction message is the first message for the
>   // key or after a state clean up.
>   if (!inputC.change) {
> return
>   }
>   // first accumulate message
>   firstRow = true
>   accumulators = function.createAccumulators()
> } else {
>   firstRow = false
> }
> {code}
> The bug has not been verified. I spotted it only by looking at the code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16589) Flink Table SQL fails/crashes with big queries with lots of fields

2020-03-13 Thread Benchao Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058745#comment-17058745
 ] 

Benchao Li commented on FLINK-16589:


[~ykt836] We did the code split only for CalcCodeGenerator in 
https://issues.apache.org/jira/browse/FLINK-15430.  

If we need to split for {{GroupAggsHandler}} too, I can help to do that.

> Flink Table SQL fails/crashes with big queries with lots of fields
> --
>
> Key: FLINK-16589
> URL: https://issues.apache.org/jira/browse/FLINK-16589
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Viet Pham
>Priority: Major
>
> Hi,
> My use case is a streaming application with a few streaming tables.
> I was trying to build a SELECT query (and registering it as a temporary view) 
> with about 200 fields/expressions out of another streaming table. The 
> application is successfully submitted to Flink cluster. However the worker 
> processes keep crashing, with the exception as quoted below. 
> It clearly mentioned in the log that this is a bug, so I fire this ticket. By 
> the way, if I lower the number of fields down to 100 then it works nicely.
> Please advice.
> Thanks a lot for all the efforts bring Flink up. It is really amazing!
> {code:java}
> java.lang.RuntimeException: Could not instantiate generated class 
> 'GroupAggsHandler$9687'at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:57)
> at 
> org.apache.flink.table.runtime.operators.aggregate.MiniBatchGroupAggFunction.open(MiniBatchGroupAggFunction.java:136)
> at 
> org.apache.flink.table.runtime.operators.bundle.AbstractMapBundleOperator.open(AbstractMapBundleOperator.java:84)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeStateAndOpen(StreamTask.java:1007)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:454)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:94)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:449)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:461)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:68)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:78)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:52)
> ... 10 moreCaused by: 
> org.apache.flink.shaded.guava18.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2203)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache.get(LocalCache.java:3937)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4739)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:66)
> ... 12 moreCaused by: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:81)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:66)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4742)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
> ... 15 moreCaused by: org.codehaus.janino.InternalCompilerException: 
> Compiling "GroupAggsHandler$9687": Code of method 
>

[GitHub] [flink] flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build ResultSubpartitionInfo and InputChannelInfo in respective constructors

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11400: [FLINK-16586][network] Build 
ResultSubpartitionInfo and InputChannelInfo in respective constructors
URL: https://github.com/apache/flink/pull/11400#issuecomment-598611898
 
 
   
   ## CI report:
   
   * a86f1e2e6785237893201dac302bc50041610d73 UNKNOWN
   * 3c660599680944e8c6f2b3b0ceaf6975b03fc5c4 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153164755) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6282)
 
   * 5216b6412ed32669643e576b1879af2122f43e4b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16589) Flink Table SQL fails/crashes with big queries with lots of fields

2020-03-13 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058742#comment-17058742
 ] 

Jingsong Lee commented on FLINK-16589:
--

If just too many fields in "select *" with "group by", it can be resolved, just 
like `ProjectionCodeGenerator`, we can extract its "loop" code generation.

> Flink Table SQL fails/crashes with big queries with lots of fields
> --
>
> Key: FLINK-16589
> URL: https://issues.apache.org/jira/browse/FLINK-16589
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Viet Pham
>Priority: Major
>
> Hi,
> My use case is a streaming application with a few streaming tables.
> I was trying to build a SELECT query (and registering it as a temporary view) 
> with about 200 fields/expressions out of another streaming table. The 
> application is successfully submitted to Flink cluster. However the worker 
> processes keep crashing, with the exception as quoted below. 
> It clearly mentioned in the log that this is a bug, so I fire this ticket. By 
> the way, if I lower the number of fields down to 100 then it works nicely.
> Please advice.
> Thanks a lot for all the efforts bring Flink up. It is really amazing!
> {code:java}
> java.lang.RuntimeException: Could not instantiate generated class 
> 'GroupAggsHandler$9687'at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:57)
> at 
> org.apache.flink.table.runtime.operators.aggregate.MiniBatchGroupAggFunction.open(MiniBatchGroupAggFunction.java:136)
> at 
> org.apache.flink.table.runtime.operators.bundle.AbstractMapBundleOperator.open(AbstractMapBundleOperator.java:84)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeStateAndOpen(StreamTask.java:1007)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:454)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$SynchronizedStreamTaskActionExecutor.runThrowing(StreamTaskActionExecutor.java:94)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:449)
> at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:461)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)at 
> java.lang.Thread.run(Thread.java:748)Caused by: 
> org.apache.flink.util.FlinkRuntimeException: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:68)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:78)
> at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:52)
> ... 10 moreCaused by: 
> org.apache.flink.shaded.guava18.com.google.common.util.concurrent.UncheckedExecutionException:
>  org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2203)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache.get(LocalCache.java:3937)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4739)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:66)
> ... 12 moreCaused by: 
> org.apache.flink.api.common.InvalidProgramException: Table program cannot be 
> compiled. This is a bug. Please file an issue.at 
> org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:81)
> at 
> org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:66)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4742)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
> at 
> org.apache.flink.shaded.guava18.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
> ... 15 moreCaused by: org.codehaus.janino.InternalCompilerException: 
> Compiling "GroupAggsHandler$9687": Code of method 
> "retract(Lorg/apache/flink/ta

[GitHub] [flink] flinkbot edited a comment on issue #7702: [FLINK-11088][Security][YARN] Allow YARN to discover pre-installed keytab files

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #7702: [FLINK-11088][Security][YARN] Allow 
YARN to discover pre-installed keytab files
URL: https://github.com/apache/flink/pull/7702#issuecomment-572195960
 
 
   
   ## CI report:
   
   * d4a17beb028586a4b8b06a2eebfe7aa21ccf4fbb UNKNOWN
   * ae58c4fc1dcbe3ea0fa70326c09803c3e5792ede Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153089193) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6253)
 
   * 0e955bd5b9f490a8ed66013ccd8ce3f512955513 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14311) Streaming File Sink end-to-end test failed on Travis

2020-03-13 Thread Yu Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058738#comment-17058738
 ] 

Yu Li commented on FLINK-14311:
---

Another instance in release-1.10 crone build: 
https://api.travis-ci.org/v3/job/661626127/log.txt

> Streaming File Sink end-to-end test failed on Travis
> 
>
> Key: FLINK-14311
> URL: https://issues.apache.org/jira/browse/FLINK-14311
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem, Tests
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Till Rohrmann
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.10.1, 1.11.0
>
>
> The {{Streaming File Sink end-to-end test}} fails on Travis because it does 
> not produce output for 10 minutes.
> https://api.travis-ci.org/v3/job/591992274/log.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11402: [FLINK-16166][build] Append additional javadoc options

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11402: [FLINK-16166][build] Append 
additional javadoc options
URL: https://github.com/apache/flink/pull/11402#issuecomment-598690738
 
 
   
   ## CI report:
   
   * 3f91303e7b54b0790f556d22da4fedc4872ac137 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153178906) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6286)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11303: [FLINK-16245] Decoupling user classloader from context classloader.

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11303: [FLINK-16245] Decoupling user 
classloader from context classloader.
URL: https://github.com/apache/flink/pull/11303#issuecomment-594377249
 
 
   
   ## CI report:
   
   * 5246383ea27ab1a98a53e58985ef5923e9e9a6c4 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153164641) Azure: 
[CANCELED](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6281)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] sjwiesman commented on issue #11367: [FLINK-16044] / [FLINK-16045] Make libraries/connectors documentation top-level sections

2020-03-13 Thread GitBox
sjwiesman commented on issue #11367: [FLINK-16044] / [FLINK-16045] Make 
libraries/connectors documentation top-level sections
URL: https://github.com/apache/flink/pull/11367#issuecomment-598721539
 
 
   I don’t know anything about seo but we’ve had issues in the past with google 
indexing the ‘stable’ branch of the docs. This means search results might 
return pages that no longer exist.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement 
new StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598702521
 
 
   
   ## CI report:
   
   * c7595ee5eb7ac60dae52bcd1368878d70d372533 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153184710) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6287)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] libenchao commented on a change in pull request #11396: [FLINK-16170][connectors/elasticsearch]SearchTemplateRequest ClassNotFoundException when use flink-sql-connector-elasticsearch7

2020-03-13 Thread GitBox
libenchao commented on a change in pull request #11396: 
[FLINK-16170][connectors/elasticsearch]SearchTemplateRequest 
ClassNotFoundException when use flink-sql-connector-elasticsearch7
URL: https://github.com/apache/flink/pull/11396#discussion_r392219528
 
 

 ##
 File path: flink-connectors/flink-sql-connector-elasticsearch7/pom.xml
 ##
 @@ -135,10 +131,6 @@ under the License.

org.elasticsearch

org.apache.flink.elasticsearch7.shaded.org.elasticsearch

-   
 
 Review comment:
   Do we need to remove this one? If yes, how about es6?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16499) Flink shaded hadoop could not work when Yarn timeline service is enabled

2020-03-13 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058717#comment-17058717
 ] 

Yang Wang commented on FLINK-16499:
---

[~chesnay] It seems that even we just use the Yarn client, we still need the 
jersey-core as dependency when the Yarn timeline service enabled.

> Flink shaded hadoop could not work when Yarn timeline service is enabled
> 
>
> Key: FLINK-16499
> URL: https://issues.apache.org/jira/browse/FLINK-16499
> Project: Flink
>  Issue Type: Bug
>  Components: BuildSystem / Shaded
>Reporter: Yang Wang
>Priority: Major
>
> When the Yarn timeline service is enabled (via 
> {{yarn.timeline-service.enabled=true}} in yarn-site.xml), flink-shaded-hadoop 
> could not work to submit Flink job to Yarn cluster. The following exception 
> will be thrown.
>  
> The root cause is the {{jersey-core-xx.jar}} is not bundled into 
> {{flink-shaded-hadoop-xx}}{{.jar}}.
>  
> {code:java}
> 2020-03-09 03:35:34,396 ERROR org.apache.flink.client.cli.CliFrontend         
>              [] - Fatal error while running command line interface.2020-03-09 
> 03:35:34,396 ERROR org.apache.flink.client.cli.CliFrontend                    
>   [] - Fatal error while running command line 
> interface.java.lang.NoClassDefFoundError: javax/ws/rs/ext/MessageBodyReader 
> at java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.defineClass(ClassLoader.java:757) ~[?:1.8.0_242] at 
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) 
> ~[?:1.8.0_242] at 
> java.net.URLClassLoader.defineClass(URLClassLoader.java:468) ~[?:1.8.0_242] 
> at java.net.URLClassLoader.access$100(URLClassLoader.java:74) ~[?:1.8.0_242] 
> at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_242] at 
> java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_242] at 
> java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_242] at 
> java.net.URLClassLoader.findClass(URLClassLoader.java:362) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.loadClass(ClassLoader.java:419) ~[?:1.8.0_242] at 
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) ~[?:1.8.0_242] 
> at java.lang.ClassLoader.loadClass(ClassLoader.java:352) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.defineClass(ClassLoader.java:757) ~[?:1.8.0_242] at 
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) 
> ~[?:1.8.0_242] at 
> java.net.URLClassLoader.defineClass(URLClassLoader.java:468) ~[?:1.8.0_242] 
> at java.net.URLClassLoader.access$100(URLClassLoader.java:74) ~[?:1.8.0_242] 
> at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_242] at 
> java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_242] at 
> java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_242] at 
> java.net.URLClassLoader.findClass(URLClassLoader.java:362) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.loadClass(ClassLoader.java:419) ~[?:1.8.0_242] at 
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) ~[?:1.8.0_242] 
> at java.lang.ClassLoader.loadClass(ClassLoader.java:352) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.defineClass(ClassLoader.java:757) ~[?:1.8.0_242] at 
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) 
> ~[?:1.8.0_242] at 
> java.net.URLClassLoader.defineClass(URLClassLoader.java:468) ~[?:1.8.0_242] 
> at java.net.URLClassLoader.access$100(URLClassLoader.java:74) ~[?:1.8.0_242] 
> at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_242] at 
> java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_242] at 
> java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_242] at 
> java.net.URLClassLoader.findClass(URLClassLoader.java:362) ~[?:1.8.0_242] at 
> java.lang.ClassLoader.loadClass(ClassLoader.java:419) ~[?:1.8.0_242] at 
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) ~[?:1.8.0_242] 
> at java.lang.ClassLoader.loadClass(ClassLoader.java:352) ~[?:1.8.0_242] at 
> org.apache.hadoop.yarn.util.timeline.TimelineUtils.(TimelineUtils.java:50)
>  ~[flink-shaded-hadoop-2-uber-2.8.3-7.0.jar:2.8.3-7.0] at 
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:179)
>  ~[flink-shaded-hadoop-2-uber-2.8.3-7.0.jar:2.8.3-7.0] at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) 
> ~[flink-shaded-hadoop-2-uber-2.8.3-7.0.jar:2.8.3-7.0] at 
> org.apache.flink.yarn.YarnClusterClientFactory.getClusterDescriptor(YarnClusterClientFactory.java:71)
>  ~[flink-dist_2.11-1.10.0-vvr-0.1-SNAPSHOT.jar:1.10.0-vvr-0.1-S

[GitHub] [flink] zentol commented on a change in pull request #11303: [FLINK-16245] Decoupling user classloader from context classloader.

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11303: [FLINK-16245] Decoupling 
user classloader from context classloader.
URL: https://github.com/apache/flink/pull/11303#discussion_r392217711
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/execution/librarycache/FlinkUserCodeClassLoaders.java
 ##
 @@ -82,4 +89,61 @@ public static ResolveOrder fromString(String resolveOrder) {
super(urls, parent);
}
}
+
+   /**
+* Ensures that holding a reference on the context class loader 
outliving the scope of user code does not prevent
+* the user classloader to be garbage collected (FLINK-16245).
+*
+* This classloader delegates to the actual user classloader. Upon 
{@link #close()}, the delegate is nulled
+* and can be garbage collected. Additional class resolution will be 
resolved solely through the bootstrap
+* classloader and most likely result in ClassNotFound exceptions.
+*/
+   private static class SafetyNetWrapperClassLoader extends URLClassLoader
+   implements Closeable {
+   private static final Logger LOG = 
LoggerFactory.getLogger(SafetyNetWrapperClassLoader.class);
+
+   private FlinkUserCodeClassLoader inner;
 
 Review comment:
   ```suggestion
private final FlinkUserCodeClassLoader inner;
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zentol commented on a change in pull request #11303: [FLINK-16245] Decoupling user classloader from context classloader.

2020-03-13 Thread GitBox
zentol commented on a change in pull request #11303: [FLINK-16245] Decoupling 
user classloader from context classloader.
URL: https://github.com/apache/flink/pull/11303#discussion_r392217711
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/execution/librarycache/FlinkUserCodeClassLoaders.java
 ##
 @@ -82,4 +89,61 @@ public static ResolveOrder fromString(String resolveOrder) {
super(urls, parent);
}
}
+
+   /**
+* Ensures that holding a reference on the context class loader 
outliving the scope of user code does not prevent
+* the user classloader to be garbage collected (FLINK-16245).
+*
+* This classloader delegates to the actual user classloader. Upon 
{@link #close()}, the delegate is nulled
+* and can be garbage collected. Additional class resolution will be 
resolved solely through the bootstrap
+* classloader and most likely result in ClassNotFound exceptions.
+*/
+   private static class SafetyNetWrapperClassLoader extends URLClassLoader
+   implements Closeable {
+   private static final Logger LOG = 
LoggerFactory.getLogger(SafetyNetWrapperClassLoader.class);
+
+   private FlinkUserCodeClassLoader inner;
 
 Review comment:
   ```suggestion
private final FlinkUserCodeClassLoader inner;
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-16554) Extract static classes from StreamTask

2020-03-13 Thread Zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijiang closed FLINK-16554.

Resolution: Fixed

Merged in master: 3b99f7e4d769f6e2b6e9fd8e36a542e063d6c369

> Extract static classes from StreamTask
> --
>
> Key: FLINK-16554
> URL: https://issues.apache.org/jira/browse/FLINK-16554
> Project: Flink
>  Issue Type: Task
>  Components: Runtime / Task
>Reporter: Roman Khachatryan
>Assignee: Roman Khachatryan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> StreamTask is currently 1400+ LOC.
> We can cut it to 1100+ by simply extracting these static classes into 
> separate files:
>  * `CheckpointingOperation`
>  * `AsyncCheckpointRunnable`



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11403: [FLINK-16316][operators] Implement 
new StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598702521
 
 
   
   ## CI report:
   
   * c7595ee5eb7ac60dae52bcd1368878d70d372533 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/153184710) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6287)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16478) add restApi to modify loglevel

2020-03-13 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16478?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058707#comment-17058707
 ] 

Yang Wang commented on FLINK-16478:
---

[~chesnay] Thanks for you response. I think it is not enough.

For k8s deployment, we could update the configmap and the changed 
log4j.properties will automatically be mount into the JM/TM. So it is enough in 
such case.

However for Yarn deployment, it is not very easy to find the location of 
log4j.properties. Also we need to login to the Yarn cluster and have enough 
permission to change the file. It is usually impossible, especially in the 
production environment.

If we could support update the log level by rest/webui, it will be very 
convenient.

> add restApi to modify loglevel 
> ---
>
> Key: FLINK-16478
> URL: https://issues.apache.org/jira/browse/FLINK-16478
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / REST
>Reporter: xiaodao
>Priority: Minor
>
> sometimes we may need to change loglevel to get more information to resolved 
> bug, now we need to stop it and modify conf/log4j.properties and resubmit it 
> ,i think it's better to add rest api to modify loglevel.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type inference for temporal table functions

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11253: [FLINK-16336][table] Add new type 
inference for temporal table functions
URL: https://github.com/apache/flink/pull/11253#issuecomment-592517598
 
 
   
   ## CI report:
   
   * d1368ad6d7262321e92a1e8c70211bc16a61dd05 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153170857) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6284)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhijiangW merged pull request #11382: [FLINK-16554][task] split StreamTask

2020-03-13 Thread GitBox
zhijiangW merged pull request #11382: [FLINK-16554][task] split StreamTask
URL: https://github.com/apache/flink/pull/11382
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] AHeise commented on a change in pull request #11303: [FLINK-16245] Decoupling user classloader from context classloader.

2020-03-13 Thread GitBox
AHeise commented on a change in pull request #11303: [FLINK-16245] Decoupling 
user classloader from context classloader.
URL: https://github.com/apache/flink/pull/11303#discussion_r392210373
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/execution/librarycache/FlinkUserCodeClassLoaders.java
 ##
 @@ -82,4 +91,78 @@ public static ResolveOrder fromString(String resolveOrder) {
super(urls, parent);
}
}
+
+   /**
+* Ensures that holding a reference on the context class loader 
outliving the scope of user code does not prevent
+* the user classloader to be garbage collected (FLINK-16245).
+*
+* This classloader delegates to the actual user classloader. Upon 
{@link #close()}, the delegate is nulled
+* and can be garbage collected. Additional class resolution will be 
resolved solely through the bootstrap
+* classloader and most likely result in ClassNotFound exceptions.
+*
+* @param  the classloader type that also needs to be closeable.
+*/
+   private static class SafetyNetWrapperClassLoader extends CloseableClassLoader
+   implements Closeable {
+   private static final Logger LOG = 
LoggerFactory.getLogger(SafetyNetWrapperClassLoader.class);
+
+   private T inner;
+
+   SafetyNetWrapperClassLoader(T inner) {
+   super(null);
+   this.inner = inner;
+   }
+
+   @Override
+   public void close() {
+   if (inner != null) {
+   try {
+   inner.close();
+   } catch (IOException e) {
+   LOG.warn("Could not close user 
classloader", e);
+   }
+   }
+   inner = null;
+   }
+
+   @Override
+   protected Class loadClass(String name, boolean resolve) 
throws ClassNotFoundException {
+   if (inner == null) {
+   return super.loadClass(name, resolve);
+   }
+
+   synchronized (getClassLoadingLock(name)) {
 
 Review comment:
   Added a common ancestor now and used delegation.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10742) Let Netty use Flink's buffers directly in credit-based mode

2020-03-13 Thread Zhijiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-10742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058701#comment-17058701
 ] 

Zhijiang commented on FLINK-10742:
--

Thanks for the kind reminder [~NicoK]. I think it is worth putting into release 
note to be aware by users. I would handle it later. 

> Let Netty use Flink's buffers directly in credit-based mode
> ---
>
> Key: FLINK-10742
> URL: https://issues.apache.org/jira/browse/FLINK-10742
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Affects Versions: 1.7.0
>Reporter: Nico Kruber
>Assignee: Yun Gao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> For credit-based flow control, we always have buffers available for data that 
> is sent to use. We could thus use them directly and not copy the network 
> stream into Netty buffers first and then into our buffers.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot commented on issue #11403: [FLINK-16316][operators] Implement new 
StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598702521
 
 
   
   ## CI report:
   
   * c7595ee5eb7ac60dae52bcd1368878d70d372533 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] pnowojski commented on a change in pull request #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
pnowojski commented on a change in pull request #11403: 
[FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for 
AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#discussion_r392202032
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/StreamOperatorBase.java
 ##
 @@ -0,0 +1,484 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.operators;
+
+import org.apache.flink.annotation.Experimental;
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.ExecutionConfig;
+import org.apache.flink.api.common.state.KeyedStateStore;
+import org.apache.flink.api.common.state.State;
+import org.apache.flink.api.common.state.StateDescriptor;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.java.functions.KeySelector;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.MetricOptions;
+import org.apache.flink.core.fs.CloseableRegistry;
+import org.apache.flink.metrics.MetricGroup;
+import org.apache.flink.runtime.checkpoint.CheckpointOptions;
+import org.apache.flink.runtime.execution.Environment;
+import org.apache.flink.runtime.jobgraph.OperatorID;
+import org.apache.flink.runtime.metrics.groups.OperatorMetricGroup;
+import org.apache.flink.runtime.metrics.groups.TaskManagerJobMetricGroup;
+import org.apache.flink.runtime.metrics.groups.UnregisteredMetricGroups;
+import org.apache.flink.runtime.state.CheckpointStreamFactory;
+import org.apache.flink.runtime.state.KeyedStateBackend;
+import org.apache.flink.runtime.state.OperatorStateBackend;
+import org.apache.flink.runtime.state.StateInitializationContext;
+import org.apache.flink.runtime.state.StateSnapshotContext;
+import org.apache.flink.runtime.state.VoidNamespace;
+import org.apache.flink.runtime.state.VoidNamespaceSerializer;
+import org.apache.flink.streaming.api.graph.StreamConfig;
+import org.apache.flink.streaming.api.watermark.Watermark;
+import org.apache.flink.streaming.runtime.streamrecord.LatencyMarker;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+import org.apache.flink.streaming.runtime.tasks.ProcessingTimeService;
+import org.apache.flink.streaming.runtime.tasks.StreamTask;
+import org.apache.flink.streaming.util.LatencyStats;
+import org.apache.flink.util.Preconditions;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.util.Arrays;
+import java.util.Locale;
+import java.util.Optional;
+
+/**
+ * New base class for all stream operators, replacing previous {@link 
AbstractStreamOperator}.
+ * Currently intended to work with {@link MultipleInputStreamOperator}.
+ *
+ * One note-able difference in comparison to {@link AbstractStreamOperator} 
is lack of
+ * {@link AbstractStreamOperator#setup(StreamTask, StreamConfig, Output)} in 
favor of initialisation
+ * in the constructor, and removed some tight coupling with classes like 
{@link StreamTask}.
+ *
+ * Methods are guaranteed not to be called concurrently.
+ *
+ * @param  The output type of the operator
+ */
+@Experimental
+public abstract class StreamOperatorBase implements StreamOperator {
 
 Review comment:
   alternative name could be `AbstractStreamOperatorV2`?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] pnowojski commented on a change in pull request #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
pnowojski commented on a change in pull request #11403: 
[FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for 
AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#discussion_r392201502
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/StreamOperatorFactoryUtil.java
 ##
 @@ -57,7 +57,13 @@
((ProcessingTimeServiceAware) 
operatorFactory).setProcessingTimeService(processingTimeService);
}
 
-   OP op = operatorFactory.createStreamOperator(containingTask, 
configuration, output);
+   // TODO: what to do with ProcessingTimeServiceAware?
+   OP op = operatorFactory.createStreamOperator(
+   new StreamOperatorInitializer<>(
+   containingTask,
+   configuration,
+   output,
+   processingTimeService));
 
 Review comment:
   @AHeise what do you think about including `processingTimeService` in 
`StreamOperatorInitializer` always, regardless of the 
`ProcessingTimeServiceAware`? Generally speaking what do you think about 
`StreamOperatorInitializer`?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
flinkbot commented on issue #11403: [FLINK-16316][operators] Implement new 
StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403#issuecomment-598697714
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit c7595ee5eb7ac60dae52bcd1368878d70d372533 (Fri Mar 13 
12:30:44 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11401: [FLINK-15991][doc][zh] Translate memory configuration docs into Chinese.

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11401: [FLINK-15991][doc][zh] Translate 
memory configuration docs into Chinese.
URL: https://github.com/apache/flink/pull/11401#issuecomment-598647884
 
 
   
   ## CI report:
   
   * 44c238715de2eb32237f37d3c38aa42f6765fd6c Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153164774) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6283)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11382: [FLINK-16554][task] split StreamTask

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11382: [FLINK-16554][task] split StreamTask
URL: https://github.com/apache/flink/pull/11382#issuecomment-597788287
 
 
   
   ## CI report:
   
   * 55525e050ada84dcd0e1ea935dffe3b7f09faa42 UNKNOWN
   * 3b898115d9c7e949cc6751b911098e9b2ee0f295 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153145930) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6278)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11402: [FLINK-16166][build] Append additional javadoc options

2020-03-13 Thread GitBox
flinkbot edited a comment on issue #11402: [FLINK-16166][build] Append 
additional javadoc options
URL: https://github.com/apache/flink/pull/11402#issuecomment-598690738
 
 
   
   ## CI report:
   
   * 3f91303e7b54b0790f556d22da4fedc4872ac137 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/153178906) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6286)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16316) Provide new base class for StreamOperators

2020-03-13 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-16316:
---
Labels: pull-request-available  (was: )

> Provide new base class for StreamOperators
> --
>
> Key: FLINK-16316
> URL: https://issues.apache.org/jira/browse/FLINK-16316
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Affects Versions: 1.10.0
>Reporter: Piotr Nowojski
>Assignee: Piotr Nowojski
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Current {{AbstractStreamOperator}} has at least couple of issues:
> # it's using deprecated {{SetupableStreamOperator}} interface, instead of 
> relaying on {{StreamOperatorFactory}}.
> # it depends on {{OneInputStreamOperator}} and {{TwoInputStreamOperator}} 
> interfaces, by providing indirectly implementation of methods like 
> {{processLatencyMarker}}, {{processLatencyMarker1}} and 
> {{processLatencyMarker2}}
> Both of those things makes {{AbstractStreamOperator}} an unsuitable base 
> implementation for {{MultipleInputStreamOperator}}. We should provide a new 
> base class (or hierarchy of base classes), without those two limitations.
> CC [~aljoscha]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] pnowojski opened a new pull request #11403: [FLINK-16316][operators] Implement new StreamOperatorBase as a replacement for AbstractStreamOperator

2020-03-13 Thread GitBox
pnowojski opened a new pull request #11403: [FLINK-16316][operators] Implement 
new StreamOperatorBase as a replacement for AbstractStreamOperator
URL: https://github.com/apache/flink/pull/11403
 
 
   Implement new StreamOperatorBase as a replacement for AbstractStreamOperator
   
   The new base class for operators tries to address couple of limitations in 
the AbstractStreamOperator like:
   - lack of support for multiple inputs
   - setup(...) method
   
   ## Brief change log
   
   Please check individual commit messages.
   
   ## Verifying this change
   
   This change is cover by existing tests and adds couple of new ones where 
necessary.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (**yes** / no)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (yes / **no** / 
don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (**yes** / no)
 - If yes, how is the feature documented? (not applicable / docs / 
**JavaDocs** / not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


  1   2   3   >