[jira] [Created] (FLINK-12610) Introduce planner rules about aggregate

2019-05-23 Thread godfrey he (JIRA)
godfrey he created FLINK-12610:
--

 Summary: Introduce planner rules about aggregate
 Key: FLINK-12610
 URL: https://issues.apache.org/jira/browse/FLINK-12610
 Project: Flink
  Issue Type: New Feature
  Components: Table SQL / Planner
Reporter: godfrey he
Assignee: godfrey he


This issue aims to introduce planner rules about aggregate, rules include:
1. {{AggregateCalcMergeRule}}, that recognizes {{Aggregate}} on top of a 
{{Calc}} and if possible aggregate through the calc or removes the calc
2. {{AggregateReduceGroupingRule}}, that reduces unless grouping columns
3. {{PruneAggregateCallRule}}, that that removes unreferenced AggregateCall 
from Aggregate
4. {{FlinkAggregateRemoveRule}}, that is copied from Calcite's 
AggregateRemoveRule, and supports SUM, MIN, MAX, AUXILIARY_GROUP aggregate 
functions in non-empty group aggregate
5. {{FlinkAggregateJoinTransposeRule}}, that is copied from Calcite's 
AggregateJoinTransposeRule, and supports Left/Right outer join and aggregate 
with AUXILIARY_GROUP



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [VOTE] Release flink-shaded 7.0, release candidate 2

2019-05-23 Thread jincheng sun
I found we should also add the `slf4j-api` as a `provided` dependency for
`flink-shaded-hadoop-2`.

@Nico, Thanks for your check, and the https://issues.apache.org/jira/browse/
FLINK-12598 is not the blocker.

@Chesnay, I remember where we already discussed this issue, the problems
caused by `provided` scope, and now, It seems nice to have added the
changes for the RC3!

What do you think? @Chesnay Schepler 

The PR can be found here: https://github.com/apache/flink-shaded/pull/64

Best,
Jincheng

Chesnay Schepler  于2019年5月23日周四 下午5:29写道:

> I found another issue:
>
>  1. jackson is bundled twice in the uber jar.
>  2. javax annotations are included in the slim jar
>
>
> As for 1), I messed up and modified the shading patterns only for
> hadoop-uber, resulting in 2 different shading patterns for jackson.
> (jackson is pulled into both jars separately; once explicitly in
> shaded-hadoop, and another time in shaded-hadoop-uber as a transitive
> dependency of avro)
>
> I've already fixed this on master.
>
> 2) happened because we didn't mark findbugs as provided like in the main
> project. I'm not quite sure why we aren't doing that, but I'm not too
> bothered by this being bundled. This kind of setups where we rely on
> Flink to provide certain dependencies are a bit brittle in general.
>
>
> On 23/05/2019 10:26, Chesnay Schepler wrote:
> > Worked for me; could you try again without parallel builds?
> >
> > Which maven version are you using?
> >
> > On 23/05/2019 09:24, Nico Kruber wrote:
> >> Thanks Jincheng,
> >> unfortunately, I think, I found a blocker:
> >>
> >> -1 (non-binding)
> >>
> >> # not working:
> >> - https://issues.apache.org/jira/browse/FLINK-12598
> >>
> >> # ok:
> >> - checked release notes
> >> - verified signatures (sha512, asc)
> >> - verified "mvn clean package -Pinclude-netty-tcnative-static"
> >>
> >> On 22/05/2019 07:32, Hequn Cheng wrote:
> >>> Thank you Jincheng for the release!
> >>>
> >>> +1 (non-binding)
> >>>
> >>> - Release notes are correct.
> >>> - Built from source archive successfully.
> >>> - Signatures and hash are correct.
> >>> - All artifacts(11 artifacts including flink-shaded) have been
> >>> deployed to
> >>> the maven central repository.
> >>>
> >>> One minor comment for the website pull request but I think it is not a
> >>> blocker.
> >>>
> >>> Best, Hequn
> >>>
> >>>
> >>> On Mon, May 20, 2019 at 9:20 AM jincheng sun  >
> >>> wrote:
> >>>
>  Hi everyone,
> 
>  Please review and vote on the release candidate #2 for the version
>  7.0, as
>  follows:
> 
>  [ ] +1, Approve the release
>  [ ] -1, Do not approve the release (please provide specific comments)
> 
>  The complete staging area is available for your review, which
>  includes:
>  * JIRA release notes [1],
>  * the official Apache source release to be deployed to
> dist.apache.org
>  [2],
>  which are signed with the key with fingerprint
>  8FEA1EE9D0048C0CCC70B7573211B0703B79EA0E [3],
>  * all artifacts to be deployed to the Maven Central Repository [4],
>  * source code tag "release-7.0-rc2" [5],
>  * website pull request listing the new release [6].
> 
>  The vote will be open for at least 72 hours. It is adopted by majority
>  approval, with at least 3 PMC affirmative votes.
> 
>  Thanks,
>  Jincheng
> 
>  [1]
> 
> 
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12345226=Html=12315522=Create_token=A5KQ-2QAV-T4JA-FDED%7C8ba061049bec0c5a72dc0191c47bb53a73b82cb4%7Clin
> 
>  [2]
> https://dist.apache.org/repos/dist/dev/flink/flink-shaded-7.0-rc2/
>  [3] https://dist.apache.org/repos/dist/release/flink/KEYS
>  [4]
> 
> https://repository.apache.org/content/repositories/orgapacheflink-1218
>  [5] https://github.com/apache/flink-shaded/tree/release-7.0-rc2
>  [6] https://github.com/apache/flink-web/pull/209
> 
> >
> >
>
>


[jira] [Created] (FLINK-12609) Align the data types of Python with Java

2019-05-23 Thread Dian Fu (JIRA)
Dian Fu created FLINK-12609:
---

 Summary: Align the data types of Python with Java
 Key: FLINK-12609
 URL: https://issues.apache.org/jira/browse/FLINK-12609
 Project: Flink
  Issue Type: Sub-task
Reporter: Dian Fu
Assignee: Dian Fu


Currently, there are some data types defined in Java not supported in Python 
such as TIMESTAMP_WITH_TIME_ZONE, TIMESTAMP_WITH_LOCAL_TIME_ZONE, etc. We 
should support them in Python once these types have been fully supported in 
Java.

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12608) Add getVertexOrThrow(ExecutionVertexID) to SchedulingTopology

2019-05-23 Thread BoWang (JIRA)
BoWang created FLINK-12608:
--

 Summary: Add getVertexOrThrow(ExecutionVertexID) to 
SchedulingTopology
 Key: FLINK-12608
 URL: https://issues.apache.org/jira/browse/FLINK-12608
 Project: Flink
  Issue Type: Sub-task
Reporter: BoWang
Assignee: BoWang






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12607) Introduce a REST API that returns the maxParallelism of a job

2019-05-23 Thread Akshay Kanfade (JIRA)
Akshay Kanfade created FLINK-12607:
--

 Summary: Introduce a REST API that returns the maxParallelism of a 
job
 Key: FLINK-12607
 URL: https://issues.apache.org/jira/browse/FLINK-12607
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / REST
Affects Versions: 1.6.3
Reporter: Akshay Kanfade


Today, Flink does not offer any way to get the maxParallelism for a job and 
it's operators through any of the REST APIs. Since, the internal state already 
tracks maxParallelism for a job and it's operators, we should expose this via 
the REST APIs so that application developer can get more insights on the 
current state.

There can be two approaches on how we can do this -

Approach 1 :

Modify the existing rest API response model to additionally expose a new field 
'maxParallelism'. Some of the REST APIs that would be affected by this
|h5. */jobs/:jobid/vertices/:vertexid*|

|h5. */jobs/:jobid*|

 

Approach 2 :

Create a new REST API that would only return maxParallelism for a job and it's 
operators.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [SURVEY] Usage of flink-ml and [DISCUSS] Delete flink-ml

2019-05-23 Thread Rong Rong
+1 for the deletion.

Also I think it also might be a good idea to update the roadmap for the
plan of removal/development since we've reached the consensus on FLIP-39.

Thanks,
Rong


On Wed, May 22, 2019 at 8:26 AM Shaoxuan Wang  wrote:

> Hi Chesnay,
> Yes, you are right. There is not any active commit planned for the legacy
> Flink-ml package. It does not matter delete it now or later. I will open a
> PR and remove it.
>
> Shaoxuan
>
> On Wed, May 22, 2019 at 7:05 PM Chesnay Schepler 
> wrote:
>
>> I believe we can remove it regardless since users could just use the 1.8
>> version against future releases.
>>
>> Generally speaking, any library/connector that is no longer actively
>> developed can be removed from the project as existing users can always
>> rely on previous versions, which should continue to work by virtue of
>> working against @Stable APIs.
>>
>> On 22/05/2019 12:08, Shaoxuan Wang wrote:
>> > Hi Flink community,
>> >
>> > We plan to delete/deprecate the legacy flink-libraries/flink-ml package
>> in
>> > Flink1.9, and replace it with the new flink-ml interface proposed in
>> FLIP39
>> > (FLINK-12470).
>> > Before we remove this package, I want to reach out to you and ask if
>> there
>> > is any active project still uses this package. Please respond to this
>> > thread and outline how you use flink-libraries/flink-ml.
>> > Depending on the replies of activity and adoption
>> > of flink-libraries/flink-ml, we will decide to either delete this
>> package
>> > in Flink1.9 or deprecate it for now & remove it in the next release
>> after
>> > 1.9.
>> >
>> > Thanks for your attention and help!
>> >
>> > Regards,
>> > Shaoxuan
>> >
>>
>>


[jira] [Created] (FLINK-12606) Move SourceFunction to flink-core

2019-05-23 Thread Aljoscha Krettek (JIRA)
Aljoscha Krettek created FLINK-12606:


 Summary: Move SourceFunction to flink-core
 Key: FLINK-12606
 URL: https://issues.apache.org/jira/browse/FLINK-12606
 Project: Flink
  Issue Type: Improvement
  Components: API / DataStream
Reporter: Aljoscha Krettek
Assignee: Aljoscha Krettek


As part of the efforts to merge the new Blink-based Table API runner 
(FLINK-11439) we need to make {{TableSource}}/{{TableSink}} interfaces 
available in {{flink-table-common}}. This requires, among other things, that 
the "physical" source interfaces are available in {{flink-core}}, independent 
of DataSet/DataStream dependencies. For {{InputFormat}} this is already the 
case, we need to do the same for {{SourceFunction}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12605) Describe new contribution process

2019-05-23 Thread Robert Metzger (JIRA)
Robert Metzger created FLINK-12605:
--

 Summary: Describe new contribution process
 Key: FLINK-12605
 URL: https://issues.apache.org/jira/browse/FLINK-12605
 Project: Flink
  Issue Type: Task
  Components: Project Website
Reporter: Robert Metzger
Assignee: Robert Metzger


The community has decided to change the contribution process to seek consensus 
in Jira first.

Update the website to reflect this change.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12604) Register TableSource/Sink as CatalogTable

2019-05-23 Thread Dawid Wysakowicz (JIRA)
Dawid Wysakowicz created FLINK-12604:


 Summary: Register TableSource/Sink as CatalogTable
 Key: FLINK-12604
 URL: https://issues.apache.org/jira/browse/FLINK-12604
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / API, Table SQL / Legacy Planner
Affects Versions: 1.9.0
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz
 Fix For: 1.9.0






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12603) Remove getOwningTaskName method from InputGate

2019-05-23 Thread zhijiang (JIRA)
zhijiang created FLINK-12603:


 Summary: Remove getOwningTaskName method from InputGate
 Key: FLINK-12603
 URL: https://issues.apache.org/jira/browse/FLINK-12603
 Project: Flink
  Issue Type: Sub-task
  Components: Runtime / Network
Reporter: zhijiang
Assignee: zhijiang


Current `InputGate#getOwningTaskName` is only used for logging in related 
components such as `BarrierBuffer`, `StreamInputProcessor`, etc. We could put 
this name in the structure of `TaskInfo`,  then the related components could 
get task name directly from `RuntimeEnvironment#getTaskInfo`.

To do so, we could simplify the interface of `InputGate`.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [DISCUSS] Correct the flink pom `artifactId` config before flink 1.9 release

2019-05-23 Thread jincheng sun
Oh, Thanks for your quick response, I have copied the content to JIRA.

Chesnay Schepler  于2019年5月23日周四 下午7:03写道:

> Please post your mail as a comment into the JIRA to consolidate the
> discussion there.
>
> On 23/05/2019 12:55, jincheng sun wrote:
> > Thank you for confirming this is an issue,and thanks a lot for your
> double
> > check. Due to the script check logic is incorrect, most of the outputs
> are
> > incorrect.
> >
> > You are right and I just pointed out the problem of the scala-free check
> > and not mentioned the final solution which I think we should discuss it.
> >
> > I tried to understand you and have modified the script a bit and have got
> > the same results as you mentioned(In my local).
> > The main changes are as follows:
> > 1) Adds test modules which should be checked:
> > !flink-fs-tests,!flink-yarn-tests,!flink-tests
> > 2) Marks the modules as infected which depend on scala trasitively or
> > depend on modules suffixed with `_{scala_version}`
> >
> > I think we should discuss the rule of how to check whether a module is
> > scala-free or not.
> > If I understand you correctly, the rule in your mind may be as follows:
> >
> > 1) All the modules should check the dependencies(excluding the
> dependencies
> > introduced by the test code). Regarding to modules flink-fs-tests,
> > flink-yarn-tests and flink-tests, the dependencies introduced by the
> > test code should also be checked.
> > 2) The checking rule is whether a module depends on scala trasitively or
> > depends on modules suffixed with `_{scala_version}`.
> > Following the above rules, we can get results you mentioned(Only 3
> modules
> > with incorrect artifact id).
> >
> > Open question:
> >
> > Currently, all the test code are also released into the repository, such
> as
> >
> http://central.maven.org/maven2/org/apache/flink/flink-avro/1.8.0/flink-avro-1.8.0-tests.jar
> > .
> > Users can also depend on these jars. My question is why we need to check
> > the test dependencies for modules flink-fs-tests, flink-yarn-tests and
> > flink-tests,
> > but not check the test dependencies for other modules?
> >
> > The solution:
> > If we follow the rule you mentioned, the change is as follows:
> > 1) Correct the check logic for the script
> > 2) Correct the artifact id for modules:  flink-connector-hive,
> > flink-queryable-state-client-java
> > 3) Add the scala dependencies for the module flink-table-api-scala due we
> > plan to add the scala code(I discussed with Timo and Aljoscha)
> >
> > If we should check other test code for other modules, maybe we need more
> > changes. But it depends on the above open question.
> >
> > I have already opened the JIRA:
> > https://issues.apache.org/jira/browse/FLINK-12602
> >
> > Feel free to discuss the solution in this mail thread or in the JIRA.
> >
> > Best,
> > Jincheng
> >
> > Chesnay Schepler  于2019年5月22日周三 下午8:50写道:
> >
> >> You do have a point, but your output is mostly incorrect.
> >>
> >> There are only 3 modules that have a suffix, which don't need it:
> >>
> >> -
> >>
> >> flink-connector-hive
> >>
> >> -
> >>
> >> flink-queryable-state-client-java
> >>
> >> -
> >>
> >> flink-table-api-scala
> >>
> >>
> >> The remaining do need it since they have dependencies with a
> scala-suffix
> >> (mostly on runtime and streaming-java).
> >>
> >> Your change does make sense to me, but there's likely another issue in
> the
> >> preceding logic that determines which module is scala-free.
> >> flink-tests for example should not be considered scala-free, since it
> >> relies on flink-runtime which contains scala and hence the scala-lang
> >> dependencies, but it apparently is given that your change detects
> something.
> >>
> >> In any case, please open a JIRA:
> >>
> >> On 22/05/2019 13:31, jincheng sun wrote:
> >>
> >> Hi all,
> >>
> >> I find a shell issue in `verify_scala_suffixes.sh`(line 145) as follows:
> >> ```
> >> grep "${module}_\d\+\.\d\+" "{}"
> >> ```
> >> This code want to find out all modules that the module's `artifactId`
> with
> >> a `scala_binary_version` suffix.
> >> but the problem is our all `artifactId` value is in the pattern of
> >> `XXX_${scala.binary.version}`, such as:
> >> ```
> >> flink-tests_${scala.binary.version}
> >> ```
> >> then the find out always empty, so this check did not take effect.
> >>
> >> When I correct the script as follows:
> >>
> >> ```
> >> grep "${module}_\\${scala.binary.version}" "{}"
> >> ```
> >> we find there more than 10 modules have incorrect `artifactId` config.
> as
> >> follows:
> >>
> >> 1.flink-connector-hive
> >> 2.flink-fs-tests
> >> 3.flink-queryable-state-client-java
> >> 4.flink-sql-connector-elasticsearch6
> >> 5.flink-sql-connector-kafka
> >> 6.flink-sql-connector-kafka-0.10
> >> 7.flink-sql-connector-kafka-0.11
> >> 8.flink-sql-connector-kafka-0.9
> >> 9.flink-table-api-scala
> >> 10.flink-tests
> >> 11.flink-yarn-tests
> >>
> >> And to fix this issue, we need a big change, such as:
> >>  

Re: [DISCUSS] Correct the flink pom `artifactId` config before flink 1.9 release

2019-05-23 Thread Chesnay Schepler
Please post your mail as a comment into the JIRA to consolidate the 
discussion there.


On 23/05/2019 12:55, jincheng sun wrote:

Thank you for confirming this is an issue,and thanks a lot for your double
check. Due to the script check logic is incorrect, most of the outputs are
incorrect.

You are right and I just pointed out the problem of the scala-free check
and not mentioned the final solution which I think we should discuss it.

I tried to understand you and have modified the script a bit and have got
the same results as you mentioned(In my local).
The main changes are as follows:
1) Adds test modules which should be checked:
!flink-fs-tests,!flink-yarn-tests,!flink-tests
2) Marks the modules as infected which depend on scala trasitively or
depend on modules suffixed with `_{scala_version}`

I think we should discuss the rule of how to check whether a module is
scala-free or not.
If I understand you correctly, the rule in your mind may be as follows:

1) All the modules should check the dependencies(excluding the dependencies
introduced by the test code). Regarding to modules flink-fs-tests,
flink-yarn-tests and flink-tests, the dependencies introduced by the
test code should also be checked.
2) The checking rule is whether a module depends on scala trasitively or
depends on modules suffixed with `_{scala_version}`.
Following the above rules, we can get results you mentioned(Only 3 modules
with incorrect artifact id).

Open question:

Currently, all the test code are also released into the repository, such as
http://central.maven.org/maven2/org/apache/flink/flink-avro/1.8.0/flink-avro-1.8.0-tests.jar
.
Users can also depend on these jars. My question is why we need to check
the test dependencies for modules flink-fs-tests, flink-yarn-tests and
flink-tests,
but not check the test dependencies for other modules?

The solution:
If we follow the rule you mentioned, the change is as follows:
1) Correct the check logic for the script
2) Correct the artifact id for modules:  flink-connector-hive,
flink-queryable-state-client-java
3) Add the scala dependencies for the module flink-table-api-scala due we
plan to add the scala code(I discussed with Timo and Aljoscha)

If we should check other test code for other modules, maybe we need more
changes. But it depends on the above open question.

I have already opened the JIRA:
https://issues.apache.org/jira/browse/FLINK-12602

Feel free to discuss the solution in this mail thread or in the JIRA.

Best,
Jincheng

Chesnay Schepler  于2019年5月22日周三 下午8:50写道:


You do have a point, but your output is mostly incorrect.

There are only 3 modules that have a suffix, which don't need it:

-

flink-connector-hive

-

flink-queryable-state-client-java

-

flink-table-api-scala


The remaining do need it since they have dependencies with a scala-suffix
(mostly on runtime and streaming-java).

Your change does make sense to me, but there's likely another issue in the
preceding logic that determines which module is scala-free.
flink-tests for example should not be considered scala-free, since it
relies on flink-runtime which contains scala and hence the scala-lang
dependencies, but it apparently is given that your change detects something.

In any case, please open a JIRA:

On 22/05/2019 13:31, jincheng sun wrote:

Hi all,

I find a shell issue in `verify_scala_suffixes.sh`(line 145) as follows:
```
grep "${module}_\d\+\.\d\+" "{}"
```
This code want to find out all modules that the module's `artifactId`  with
a `scala_binary_version` suffix.
but the problem is our all `artifactId` value is in the pattern of
`XXX_${scala.binary.version}`, such as:
```
flink-tests_${scala.binary.version}
```
then the find out always empty, so this check did not take effect.

When I correct the script as follows:

```
grep "${module}_\\${scala.binary.version}" "{}"
```
we find there more than 10 modules have incorrect `artifactId` config. as
follows:

1.flink-connector-hive
2.flink-fs-tests
3.flink-queryable-state-client-java
4.flink-sql-connector-elasticsearch6
5.flink-sql-connector-kafka
6.flink-sql-connector-kafka-0.10
7.flink-sql-connector-kafka-0.11
8.flink-sql-connector-kafka-0.9
9.flink-table-api-scala
10.flink-tests
11.flink-yarn-tests

And to fix this issue, we need a big change, such as:
 - correct the `artifactId` config.
 - update the dependency of relation modules.
 - release the connector into the repo.
 - update some of the doc for `connector.zh.md` and `connector.md`.
 - others

>From the points of my view, It's better to fix this issue before the flink
1.9 release.

What do you think?

NOTE: Please remind me if I have a missing above!

1. The script code 
change:https://github.com/sunjincheng121/flink/commit/736f16a8a76aaef1018cc754f0effec119e43120
2. The CI test result: https://api.travis-ci.org/v3/job/535719615/log.txt

Regards,
Jincheng








Re: [DISCUSS] Correct the flink pom `artifactId` config before flink 1.9 release

2019-05-23 Thread jincheng sun
Thank you for confirming this is an issue,and thanks a lot for your double
check. Due to the script check logic is incorrect, most of the outputs are
incorrect.

You are right and I just pointed out the problem of the scala-free check
and not mentioned the final solution which I think we should discuss it.

I tried to understand you and have modified the script a bit and have got
the same results as you mentioned(In my local).
The main changes are as follows:
1) Adds test modules which should be checked:
!flink-fs-tests,!flink-yarn-tests,!flink-tests
2) Marks the modules as infected which depend on scala trasitively or
depend on modules suffixed with `_{scala_version}`

I think we should discuss the rule of how to check whether a module is
scala-free or not.
If I understand you correctly, the rule in your mind may be as follows:

1) All the modules should check the dependencies(excluding the dependencies
introduced by the test code). Regarding to modules flink-fs-tests,
   flink-yarn-tests and flink-tests, the dependencies introduced by the
test code should also be checked.
2) The checking rule is whether a module depends on scala trasitively or
depends on modules suffixed with `_{scala_version}`.
Following the above rules, we can get results you mentioned(Only 3 modules
with incorrect artifact id).

Open question:

Currently, all the test code are also released into the repository, such as
http://central.maven.org/maven2/org/apache/flink/flink-avro/1.8.0/flink-avro-1.8.0-tests.jar
.
Users can also depend on these jars. My question is why we need to check
the test dependencies for modules flink-fs-tests, flink-yarn-tests and
flink-tests,
but not check the test dependencies for other modules?

The solution:
If we follow the rule you mentioned, the change is as follows:
1) Correct the check logic for the script
2) Correct the artifact id for modules:  flink-connector-hive,
flink-queryable-state-client-java
3) Add the scala dependencies for the module flink-table-api-scala due we
plan to add the scala code(I discussed with Timo and Aljoscha)

If we should check other test code for other modules, maybe we need more
changes. But it depends on the above open question.

I have already opened the JIRA:
https://issues.apache.org/jira/browse/FLINK-12602

Feel free to discuss the solution in this mail thread or in the JIRA.

Best,
Jincheng

Chesnay Schepler  于2019年5月22日周三 下午8:50写道:

> You do have a point, but your output is mostly incorrect.
>
> There are only 3 modules that have a suffix, which don't need it:
>
>-
>
>flink-connector-hive
>
>-
>
>flink-queryable-state-client-java
>
>-
>
>flink-table-api-scala
>
>
> The remaining do need it since they have dependencies with a scala-suffix
> (mostly on runtime and streaming-java).
>
> Your change does make sense to me, but there's likely another issue in the
> preceding logic that determines which module is scala-free.
> flink-tests for example should not be considered scala-free, since it
> relies on flink-runtime which contains scala and hence the scala-lang
> dependencies, but it apparently is given that your change detects something.
>
> In any case, please open a JIRA:
>
> On 22/05/2019 13:31, jincheng sun wrote:
>
> Hi all,
>
> I find a shell issue in `verify_scala_suffixes.sh`(line 145) as follows:
> ```
> grep "${module}_\d\+\.\d\+" "{}"
> ```
> This code want to find out all modules that the module's `artifactId`  with
> a `scala_binary_version` suffix.
> but the problem is our all `artifactId` value is in the pattern of
> `XXX_${scala.binary.version}`, such as:
> ```
> flink-tests_${scala.binary.version}
> ```
> then the find out always empty, so this check did not take effect.
>
> When I correct the script as follows:
>
> ```
> grep "${module}_\\${scala.binary.version}" "{}"
> ```
> we find there more than 10 modules have incorrect `artifactId` config. as
> follows:
>
> 1.flink-connector-hive
> 2.flink-fs-tests
> 3.flink-queryable-state-client-java
> 4.flink-sql-connector-elasticsearch6
> 5.flink-sql-connector-kafka
> 6.flink-sql-connector-kafka-0.10
> 7.flink-sql-connector-kafka-0.11
> 8.flink-sql-connector-kafka-0.9
> 9.flink-table-api-scala
> 10.flink-tests
> 11.flink-yarn-tests
>
> And to fix this issue, we need a big change, such as:
> - correct the `artifactId` config.
> - update the dependency of relation modules.
> - release the connector into the repo.
> - update some of the doc for `connector.zh.md` and `connector.md`.
> - others
>
> >From the points of my view, It's better to fix this issue before the flink
> 1.9 release.
>
> What do you think?
>
> NOTE: Please remind me if I have a missing above!
>
> 1. The script code 
> change:https://github.com/sunjincheng121/flink/commit/736f16a8a76aaef1018cc754f0effec119e43120
> 2. The CI test result: https://api.travis-ci.org/v3/job/535719615/log.txt
>
> Regards,
> Jincheng
>
>
>
>


[jira] [Created] (FLINK-12602) Correct the flink pom `artifactId` config and scala-free check logic

2019-05-23 Thread sunjincheng (JIRA)
sunjincheng created FLINK-12602:
---

 Summary: Correct the flink pom `artifactId` config and scala-free 
check logic
 Key: FLINK-12602
 URL: https://issues.apache.org/jira/browse/FLINK-12602
 Project: Flink
  Issue Type: Bug
  Components: Build System
Affects Versions: 1.9.0
Reporter: sunjincheng


I find a shell issue in `verify_scala_suffixes.sh`(line 145) as follows:
{code}
grep "${module}_\d\+\.\d\+" "{}"
{code}
This code want to find out all modules that the module's `artifactId`  with a 
`scala_binary_version` suffix. 
but the problem is our all `artifactId` value is in the pattern of 
`XXX_${scala.binary.version}`, such as:
{code}
flink-tests_${scala.binary.version}
{code}
then the result always empty, so this check did not take effect.

I have already initiated a discussion of the issue. Please check the Mail 
thread here for details.
http://mail-archives.apache.org/mod_mbox/flink-dev/201905.mbox/%3ccajsjtkw+8mcsc0fvneyaovl_ttrr_uuosx-tfgxj5gfqp1a...@mail.gmail.com%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12601) Register DataStream/DataSet as DataStream/SetTableOperations in Catalog

2019-05-23 Thread Dawid Wysakowicz (JIRA)
Dawid Wysakowicz created FLINK-12601:


 Summary: Register DataStream/DataSet as 
DataStream/SetTableOperations in Catalog
 Key: FLINK-12601
 URL: https://issues.apache.org/jira/browse/FLINK-12601
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / API, Table SQL / Legacy Planner
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12600) Introduce planner rules to do deterministic rewriting on RelNode

2019-05-23 Thread godfrey he (JIRA)
godfrey he created FLINK-12600:
--

 Summary: Introduce planner rules to do deterministic rewriting on 
RelNode 
 Key: FLINK-12600
 URL: https://issues.apache.org/jira/browse/FLINK-12600
 Project: Flink
  Issue Type: New Feature
  Components: Table SQL / Planner
Reporter: godfrey he
Assignee: godfrey he


This issue aims to introduce planner rules to to do deterministic rewriting on 
RelNode , rules include:
1. {{FlinkLimit0RemoveRule}} that rewrites `limit 0` to empty {{Values}}
2. {{FlinkRewriteSubQueryRule}} that rewrites a {{Filter}} with condition: 
`(select count(*) from T) > 0` to a {{Filter}} with condition: `exists(select * 
from T)`
3. {{ReplaceIntersectWithSemiJoinRule}} that rewrites distinct {{Intersect}} to 
a distinct {{Aggregate}} on a SEMI {{Join}}.
4. {{ReplaceMinusWithAntiJoinRule}} that rewrite distinct {{Minus}} to a 
distinct {{Aggregate}} on an ANTI {{Join}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [VOTE] Release flink-shaded 7.0, release candidate 2

2019-05-23 Thread Chesnay Schepler

I found another issue:

1. jackson is bundled twice in the uber jar.
2. javax annotations are included in the slim jar


As for 1), I messed up and modified the shading patterns only for 
hadoop-uber, resulting in 2 different shading patterns for jackson.
(jackson is pulled into both jars separately; once explicitly in 
shaded-hadoop, and another time in shaded-hadoop-uber as a transitive 
dependency of avro)


I've already fixed this on master.

2) happened because we didn't mark findbugs as provided like in the main 
project. I'm not quite sure why we aren't doing that, but I'm not too 
bothered by this being bundled. This kind of setups where we rely on 
Flink to provide certain dependencies are a bit brittle in general.



On 23/05/2019 10:26, Chesnay Schepler wrote:

Worked for me; could you try again without parallel builds?

Which maven version are you using?

On 23/05/2019 09:24, Nico Kruber wrote:

Thanks Jincheng,
unfortunately, I think, I found a blocker:

-1 (non-binding)

# not working:
- https://issues.apache.org/jira/browse/FLINK-12598

# ok:
- checked release notes
- verified signatures (sha512, asc)
- verified "mvn clean package -Pinclude-netty-tcnative-static"

On 22/05/2019 07:32, Hequn Cheng wrote:

Thank you Jincheng for the release!

+1 (non-binding)

- Release notes are correct.
- Built from source archive successfully.
- Signatures and hash are correct.
- All artifacts(11 artifacts including flink-shaded) have been 
deployed to

the maven central repository.

One minor comment for the website pull request but I think it is not a
blocker.

Best, Hequn


On Mon, May 20, 2019 at 9:20 AM jincheng sun 
wrote:


Hi everyone,

Please review and vote on the release candidate #2 for the version 
7.0, as

follows:

[ ] +1, Approve the release
[ ] -1, Do not approve the release (please provide specific comments)

The complete staging area is available for your review, which 
includes:

* JIRA release notes [1],
* the official Apache source release to be deployed to dist.apache.org
[2],
which are signed with the key with fingerprint
8FEA1EE9D0048C0CCC70B7573211B0703B79EA0E [3],
* all artifacts to be deployed to the Maven Central Repository [4],
* source code tag "release-7.0-rc2" [5],
* website pull request listing the new release [6].

The vote will be open for at least 72 hours. It is adopted by majority
approval, with at least 3 PMC affirmative votes.

Thanks,
Jincheng

[1]

https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12345226=Html=12315522=Create_token=A5KQ-2QAV-T4JA-FDED%7C8ba061049bec0c5a72dc0191c47bb53a73b82cb4%7Clin 


[2] https://dist.apache.org/repos/dist/dev/flink/flink-shaded-7.0-rc2/
[3] https://dist.apache.org/repos/dist/release/flink/KEYS
[4] 
https://repository.apache.org/content/repositories/orgapacheflink-1218

[5] https://github.com/apache/flink-shaded/tree/release-7.0-rc2
[6] https://github.com/apache/flink-web/pull/209








Re: [VOTE] Release flink-shaded 7.0, release candidate 2

2019-05-23 Thread Chesnay Schepler

Worked for me; could you try again without parallel builds?

Which maven version are you using?

On 23/05/2019 09:24, Nico Kruber wrote:

Thanks Jincheng,
unfortunately, I think, I found a blocker:

-1 (non-binding)

# not working:
- https://issues.apache.org/jira/browse/FLINK-12598

# ok:
- checked release notes
- verified signatures (sha512, asc)
- verified "mvn clean package -Pinclude-netty-tcnative-static"

On 22/05/2019 07:32, Hequn Cheng wrote:

Thank you Jincheng for the release!

+1 (non-binding)

- Release notes are correct.
- Built from source archive successfully.
- Signatures and hash are correct.
- All artifacts(11 artifacts including flink-shaded) have been deployed to
the maven central repository.

One minor comment for the website pull request but I think it is not a
blocker.

Best, Hequn


On Mon, May 20, 2019 at 9:20 AM jincheng sun 
wrote:


Hi everyone,

Please review and vote on the release candidate #2 for the version 7.0, as
follows:

[ ] +1, Approve the release
[ ] -1, Do not approve the release (please provide specific comments)

The complete staging area is available for your review, which includes:
* JIRA release notes [1],
* the official Apache source release to be deployed to dist.apache.org
[2],
which are signed with the key with fingerprint
8FEA1EE9D0048C0CCC70B7573211B0703B79EA0E [3],
* all artifacts to be deployed to the Maven Central Repository [4],
* source code tag "release-7.0-rc2" [5],
* website pull request listing the new release [6].

The vote will be open for at least 72 hours. It is adopted by majority
approval, with at least 3 PMC affirmative votes.

Thanks,
Jincheng

[1]

https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12345226=Html=12315522=Create_token=A5KQ-2QAV-T4JA-FDED%7C8ba061049bec0c5a72dc0191c47bb53a73b82cb4%7Clin
[2] https://dist.apache.org/repos/dist/dev/flink/flink-shaded-7.0-rc2/
[3] https://dist.apache.org/repos/dist/release/flink/KEYS
[4] https://repository.apache.org/content/repositories/orgapacheflink-1218
[5] https://github.com/apache/flink-shaded/tree/release-7.0-rc2
[6] https://github.com/apache/flink-web/pull/209





Re: [VOTE] Release flink-shaded 7.0, release candidate 2

2019-05-23 Thread Nico Kruber
Thanks Jincheng,
unfortunately, I think, I found a blocker:

-1 (non-binding)

# not working:
- https://issues.apache.org/jira/browse/FLINK-12598

# ok:
- checked release notes
- verified signatures (sha512, asc)
- verified "mvn clean package -Pinclude-netty-tcnative-static"

On 22/05/2019 07:32, Hequn Cheng wrote:
> Thank you Jincheng for the release!
> 
> +1 (non-binding)
> 
> - Release notes are correct.
> - Built from source archive successfully.
> - Signatures and hash are correct.
> - All artifacts(11 artifacts including flink-shaded) have been deployed to
> the maven central repository.
> 
> One minor comment for the website pull request but I think it is not a
> blocker.
> 
> Best, Hequn
> 
> 
> On Mon, May 20, 2019 at 9:20 AM jincheng sun 
> wrote:
> 
>> Hi everyone,
>>
>> Please review and vote on the release candidate #2 for the version 7.0, as
>> follows:
>>
>> [ ] +1, Approve the release
>> [ ] -1, Do not approve the release (please provide specific comments)
>>
>> The complete staging area is available for your review, which includes:
>> * JIRA release notes [1],
>> * the official Apache source release to be deployed to dist.apache.org
>> [2],
>> which are signed with the key with fingerprint
>> 8FEA1EE9D0048C0CCC70B7573211B0703B79EA0E [3],
>> * all artifacts to be deployed to the Maven Central Repository [4],
>> * source code tag "release-7.0-rc2" [5],
>> * website pull request listing the new release [6].
>>
>> The vote will be open for at least 72 hours. It is adopted by majority
>> approval, with at least 3 PMC affirmative votes.
>>
>> Thanks,
>> Jincheng
>>
>> [1]
>>
>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12345226=Html=12315522=Create_token=A5KQ-2QAV-T4JA-FDED%7C8ba061049bec0c5a72dc0191c47bb53a73b82cb4%7Clin
>> [2] https://dist.apache.org/repos/dist/dev/flink/flink-shaded-7.0-rc2/
>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
>> [4] https://repository.apache.org/content/repositories/orgapacheflink-1218
>> [5] https://github.com/apache/flink-shaded/tree/release-7.0-rc2
>> [6] https://github.com/apache/flink-web/pull/209
>>
> 

-- 
Nico Kruber | Solutions Architect
Follow us @VervericaData Ververica
--
Join Flink Forward - The Apache Flink Conference
Stream Processing | Event Driven | Real Time
--
Data Artisans GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
--
Data Artisans GmbH
Registered at Amtsgericht Charlottenburg: HRB 158244 B
Managing Directors: Dr. Kostas Tzoumas, Dr. Stephan Ewen



signature.asc
Description: OpenPGP digital signature


[jira] [Created] (FLINK-12599) Add Jobgraph zooming with mouse wheel

2019-05-23 Thread Nico Kruber (JIRA)
Nico Kruber created FLINK-12599:
---

 Summary: Add Jobgraph zooming with mouse wheel
 Key: FLINK-12599
 URL: https://issues.apache.org/jira/browse/FLINK-12599
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Web Frontend
Affects Versions: 1.9.0
Reporter: Nico Kruber


The new web UI looks nice, but I almost didn't find a way back after zooming in:

If I click twice into the job graph of a running job, the graph zooms in. After 
clicking multiple times, it got too large and I wanted to zoom out:
 * the zoom slider was not clearly visible (blue on a blue operator behind it)
 * the mouse wheel did not zoom (neither in nor out)

Can we make this a bit more visible / accessible?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12598) Flink-shaded's shade-sources profile does not work anymore

2019-05-23 Thread Nico Kruber (JIRA)
Nico Kruber created FLINK-12598:
---

 Summary: Flink-shaded's shade-sources profile does not work anymore
 Key: FLINK-12598
 URL: https://issues.apache.org/jira/browse/FLINK-12598
 Project: Flink
  Issue Type: Bug
  Components: BuildSystem / Shaded
Affects Versions: shaded-7.0
Reporter: Nico Kruber
 Fix For: shaded-7.0


{code}
> mvn clean package -Pshade-sources
...
[INFO] --- maven-shade-plugin:3.0.0:shade (shade-flink) @ flink-shaded-hadoop-2 
---
[INFO] Excluding org.apache.commons:commons-compress:jar:1.18 from the shaded 
jar.
[INFO] Excluding org.apache.avro:avro:jar:1.8.2 from the shaded jar.
[INFO] Including org.codehaus.jackson:jackson-core-asl:jar:1.9.13 in the shaded 
jar.
[INFO] Including org.codehaus.jackson:jackson-mapper-asl:jar:1.9.13 in the 
shaded jar.
[INFO] Excluding com.thoughtworks.paranamer:paranamer:jar:2.7 from the shaded 
jar.
[INFO] Excluding org.xerial.snappy:snappy-java:jar:1.1.4 from the shaded jar.
[INFO] Excluding org.tukaani:xz:jar:1.5 from the shaded jar.
[INFO] Excluding org.slf4j:slf4j-api:jar:1.7.7 from the shaded jar.
[INFO] Including org.apache.hadoop:hadoop-common:jar:2.4.1 in the shaded jar.
[INFO] Including org.apache.hadoop:hadoop-annotations:jar:2.4.1 in the shaded 
jar.
[INFO] Including com.google.guava:guava:jar:11.0.2 in the shaded jar.
[INFO] Excluding commons-cli:commons-cli:jar:1.3.1 from the shaded jar.
[INFO] Excluding org.apache.commons:commons-math3:jar:3.5 from the shaded jar.
[INFO] Excluding xmlenc:xmlenc:jar:0.52 from the shaded jar.
[INFO] Including commons-httpclient:commons-httpclient:jar:3.1 in the shaded 
jar.
[INFO] Excluding commons-codec:commons-codec:jar:1.10 from the shaded jar.
[INFO] Excluding commons-io:commons-io:jar:2.4 from the shaded jar.
[INFO] Excluding commons-net:commons-net:jar:3.1 from the shaded jar.
[INFO] Excluding commons-collections:commons-collections:jar:3.2.2 from the 
shaded jar.
[INFO] Excluding javax.servlet:servlet-api:jar:2.5 from the shaded jar.
[INFO] Excluding commons-el:commons-el:jar:1.0 from the shaded jar.
[INFO] Excluding commons-logging:commons-logging:jar:1.1.3 from the shaded jar.
[INFO] Excluding log4j:log4j:jar:1.2.17 from the shaded jar.
[INFO] Including net.java.dev.jets3t:jets3t:jar:0.9.0 in the shaded jar.
[INFO] Including org.apache.httpcomponents:httpclient:jar:4.5.3 in the shaded 
jar.
[INFO] Including org.apache.httpcomponents:httpcore:jar:4.4.6 in the shaded jar.
[INFO] Excluding com.jamesmurty.utils:java-xmlbuilder:jar:0.4 from the shaded 
jar.
[INFO] Excluding commons-lang:commons-lang:jar:2.6 from the shaded jar.
[INFO] Excluding commons-configuration:commons-configuration:jar:1.7 from the 
shaded jar.
[INFO] Excluding commons-digester:commons-digester:jar:1.8.1 from the shaded 
jar.
[INFO] Excluding org.slf4j:slf4j-log4j12:jar:1.7.15 from the shaded jar.
[INFO] Including com.google.protobuf:protobuf-java:jar:2.5.0 in the shaded jar.
[INFO] Including org.apache.hadoop:hadoop-auth:jar:2.4.1 in the shaded jar.
[INFO] Excluding com.jcraft:jsch:jar:0.1.42 from the shaded jar.
[INFO] Including com.google.code.findbugs:jsr305:jar:1.3.9 in the shaded jar.
[WARNING] Could not get sources for 
com.google.code.findbugs:jsr305:jar:1.3.9:compile
[INFO] 
[INFO] Reactor Summary:
[INFO] 
[INFO] flink-shaded 7.0 ... SUCCESS [  0.771 s]
[INFO] flink-shaded-force-shading 7.0 . SUCCESS [  0.951 s]
[INFO] flink-shaded-asm-6 6.2.1-7.0 ... SUCCESS [  1.469 s]
[INFO] flink-shaded-guava-18 18.0-7.0 . SKIPPED
[INFO] flink-shaded-netty-4 4.1.32.Final-7.0 .. SKIPPED
[INFO] flink-shaded-netty-tcnative-dynamic 2.0.25.Final-7.0 SUCCESS [  2.195 s]
[INFO] flink-shaded-jackson-parent 2.9.8-7.0 .. SUCCESS [  0.161 s]
[INFO] flink-shaded-jackson-2 2.9.8-7.0 ... SKIPPED
[INFO] flink-shaded-jackson-module-jsonSchema-2 2.9.8-7.0 . SKIPPED
[INFO] flink-shaded-hadoop-2 2.4.1-7.0  FAILURE [  2.597 s]
[INFO] flink-shaded-hadoop-2-uber 2.4.1-7.0 ... SKIPPED
[INFO] 
[INFO] BUILD FAILURE
[INFO] 
[INFO] Total time:  3.499 s (Wall Clock)
[INFO] Finished at: 2019-05-23T09:22:04+02:00
[INFO] 
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-shade-plugin:3.0.0:shade (shade-flink) on 
project flink-shaded-hadoop-2: Execution shade-flink of goal 
org.apache.maven.plugins:maven-shade-plugin:3.0.0:shade failed.: 
NullPointerException -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run 

[jira] [Created] (FLINK-12597) Remove the legacy flink-libraries/flink-ml

2019-05-23 Thread Shaoxuan Wang (JIRA)
Shaoxuan Wang created FLINK-12597:
-

 Summary: Remove the legacy flink-libraries/flink-ml
 Key: FLINK-12597
 URL: https://issues.apache.org/jira/browse/FLINK-12597
 Project: Flink
  Issue Type: Sub-task
  Components: Library / Machine Learning
Affects Versions: 1.9.0
Reporter: Shaoxuan Wang
Assignee: Luo Gen


As discussed in dev-ml, we decided to delete the legacy 
flink-libraries/flink-ml, so as to the flink-libraries/flink-ml-uber. There is 
not any further development planned for this legacy flink-ml package in 1.9 or 
even future. Users could just use the 1.8 version if their products/projects 
still rely on this package.

[1] 
[http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/SURVEY-Usage-of-flink-ml-and-DISCUSS-Delete-flink-ml-td29057.html]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12596) Unify all kinds of cluster into ClusterClient in FlinkShell

2019-05-23 Thread Jeff Zhang (JIRA)
Jeff Zhang created FLINK-12596:
--

 Summary: Unify all kinds of cluster into ClusterClient in 
FlinkShell
 Key: FLINK-12596
 URL: https://issues.apache.org/jira/browse/FLINK-12596
 Project: Flink
  Issue Type: Improvement
  Components: Scala Shell
Affects Versions: 1.8.0
Reporter: Jeff Zhang






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-12595) KinesisDataFetcherTest.testOriginalExceptionIsPreservedWhenInterruptedDuringShutdown deadlocks

2019-05-23 Thread Dawid Wysakowicz (JIRA)
Dawid Wysakowicz created FLINK-12595:


 Summary: 
KinesisDataFetcherTest.testOriginalExceptionIsPreservedWhenInterruptedDuringShutdown
 deadlocks
 Key: FLINK-12595
 URL: https://issues.apache.org/jira/browse/FLINK-12595
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kinesis, Tests
Affects Versions: 1.9.0
Reporter: Dawid Wysakowicz


https://api.travis-ci.org/v3/job/535738122/log.txt





--
This message was sent by Atlassian JIRA
(v7.6.3#76005)