[jira] [Commented] (FLINK-20771) Hive partition is not added when there is a lot of data

2020-12-26 Thread hehuiyuan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17254980#comment-17254980
 ] 

hehuiyuan commented on FLINK-20771:
---

Hi  [~ZhuShang] , fix , it is good.

> Hive partition is not added when there is a lot of data
> ---
>
> Key: FLINK-20771
> URL: https://issues.apache.org/jira/browse/FLINK-20771
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Reporter: hehuiyuan
>Priority: Major
> Attachments: image-2020-12-25-18-09-42-707.png, 
> image-2020-12-25-18-15-07-519.png, image-2020-12-25-18-19-53-746.png
>
>
> Hive partition is not added when the data is huge .
> !image-2020-12-25-18-09-42-707.png|width=437,height=115!
>   Before partition commit, *inProgressPart* will be reinitialize .
> the current bucket is active , so the  notifyBucketInactive is  not executed.
> !image-2020-12-25-18-15-07-519.png|width=574,height=192!
>  
> !image-2020-12-25-18-19-53-746.png|width=697,height=56!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-20427) Remove CheckpointConfig.setPreferCheckpointForRecovery because it can lead to data loss

2020-12-26 Thread Liwen Liu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20427?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17254988#comment-17254988
 ] 

Liwen Liu commented on FLINK-20427:
---

Hi,[~trohrmann]

Could you assign this ticket to me ?

Thanks~

> Remove CheckpointConfig.setPreferCheckpointForRecovery because it can lead to 
> data loss
> ---
>
> Key: FLINK-20427
> URL: https://issues.apache.org/jira/browse/FLINK-20427
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream, Runtime / Checkpointing
>Affects Versions: 1.12.0
>Reporter: Till Rohrmann
>Priority: Critical
> Fix For: 1.13.0
>
>
> The {{CheckpointConfig.setPreferCheckpointForRecovery}} allows to configure 
> whether Flink prefers checkpoints for recovery if the 
> {{CompletedCheckpointStore}} contains savepoints and checkpoints. This is 
> problematic because due to this feature, Flink might prefer older checkpoints 
> over newer savepoints for recovery. Since some components expect that the 
> always the latest checkpoint/savepoint is used (e.g. the 
> {{SourceCoordinator}}), it breaks assumptions and can lead to 
> {{SourceSplits}} which are not read. This effectively means that the system 
> loses data. Similarly, this behaviour can cause that exactly once sinks might 
> output results multiple times which violates the processing guarantees. 
> Hence, I believe that we should remove this setting because it changes 
> Flink's behaviour in some very significant way potentially w/o the user 
> noticing.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-training] sv3ndk opened a new pull request #15: add support for Flink 1.12

2020-12-26 Thread GitBox


sv3ndk opened a new pull request #15:
URL: https://github.com/apache/flink-training/pull/15


   hi,
   
   While reading Flink documentation, I noticed that 
[learn-flink/etl.html](https://ci.apache.org/projects/flink/flink-docs-release-1.12/learn-flink/etl.html)
 has a link at the bottom of the page to 
https://github.com/apache/flink-training/tree/release-1.12/rides-and-fares, 
which does not exists yet.
   
   I tried to bump up the Flink version here to 1.12 and everything seems to 
work (I didn't do all the exercises though).
   
   Based on the gradle warning I saw during `./gradlew test shadowJar`, I also 
removed the following statements from the scala exercises: 
   ```
env.setStreamTimeCharacteristic(TimeCharacteristic.EventTime)
   ```
   since it's my understanding that this method is now deprecated and 
event-time characteristics is the default anyway (cf 
https://github.com/apache/flink/blob/release-1.12/flink-streaming-scala/src/main/scala/org/apache/flink/streaming/api/scala/StreamExecutionEnvironment.scala#L411
 ).
   
   Hope this helps, thanks for making this training freely available :)
   
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14465: [FLINK-20594][runtime] Remove DefaultExecutionSlotAllocator and its Factory class.

2020-12-26 Thread GitBox


flinkbot edited a comment on pull request #14465:
URL: https://github.com/apache/flink/pull/14465#issuecomment-749579707


   
   ## CI report:
   
   * 2147c4b63bfdf589e2c5ab167b0c41cacc7a514c Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11260)
 
   * 23d374243c9726e72ec4b1e1e87739739cded4b9 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14465: [FLINK-20594][runtime] Remove DefaultExecutionSlotAllocator and its Factory class.

2020-12-26 Thread GitBox


flinkbot edited a comment on pull request #14465:
URL: https://github.com/apache/flink/pull/14465#issuecomment-749579707


   
   ## CI report:
   
   * 2147c4b63bfdf589e2c5ab167b0c41cacc7a514c Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11260)
 
   * 23d374243c9726e72ec4b1e1e87739739cded4b9 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11342)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] sv3ndk opened a new pull request #14497: Typo doc operation playground

2020-12-26 Thread GitBox


sv3ndk opened a new pull request #14497:
URL: https://github.com/apache/flink/pull/14497


   
   
   ## What is the purpose of the change
   
   fix typo in "Flink Operation Playground": "are generate" => "are generated"
   
   ## Brief change log
   
   fix typo
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable 
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #14497: Typo doc operation playground

2020-12-26 Thread GitBox


flinkbot commented on pull request #14497:
URL: https://github.com/apache/flink/pull/14497#issuecomment-751373419


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 1d7173eb96b45442008011d6008f7e7a3831b116 (Sat Dec 26 
16:46:00 UTC 2020)
   
   **Warnings:**
* Documentation files were touched, but no `.zh.md` files: Update Chinese 
documentation or file Jira ticket.
* **Invalid pull request title: No valid Jira ID provided**
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #14497: Typo doc operation playground

2020-12-26 Thread GitBox


flinkbot commented on pull request #14497:
URL: https://github.com/apache/flink/pull/14497#issuecomment-751375809


   
   ## CI report:
   
   * 1d7173eb96b45442008011d6008f7e7a3831b116 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14497: Typo doc operation playground

2020-12-26 Thread GitBox


flinkbot edited a comment on pull request #14497:
URL: https://github.com/apache/flink/pull/14497#issuecomment-751375809


   
   ## CI report:
   
   * 1d7173eb96b45442008011d6008f7e7a3831b116 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11343)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14497: Typo doc operation playground

2020-12-26 Thread GitBox


flinkbot edited a comment on pull request #14497:
URL: https://github.com/apache/flink/pull/14497#issuecomment-751375809


   
   ## CI report:
   
   * 1d7173eb96b45442008011d6008f7e7a3831b116 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11343)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14465: [FLINK-20594][runtime] Remove DefaultExecutionSlotAllocator and its Factory class.

2020-12-26 Thread GitBox


flinkbot edited a comment on pull request #14465:
URL: https://github.com/apache/flink/pull/14465#issuecomment-749579707


   
   ## CI report:
   
   * 23d374243c9726e72ec4b1e1e87739739cded4b9 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11342)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-20461) YARNFileReplicationITCase.testPerJobModeWithDefaultFileReplication

2020-12-26 Thread Zhenqiu Huang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20461?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17255156#comment-17255156
 ] 

Zhenqiu Huang commented on FLINK-20461:
---

[~xintongsong] [~hxbks2ks]
Tried run the test class 100 times in IntelliJ, none of them failed. I will try 
to change the azure-pipelines to run/repeat this single test.



> YARNFileReplicationITCase.testPerJobModeWithDefaultFileReplication
> --
>
> Key: FLINK-20461
> URL: https://issues.apache.org/jira/browse/FLINK-20461
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.12.0
>Reporter: Huang Xingbo
>Assignee: Zhenqiu Huang
>Priority: Major
>  Labels: testability
>
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10450&view=logs&j=fc5181b0-e452-5c8f-68de-1097947f6483&t=62110053-334f-5295-a0ab-80dd7e2babbf]
> {code:java}
> [ERROR] 
> testPerJobModeWithDefaultFileReplication(org.apache.flink.yarn.YARNFileReplicationITCase)
>  Time elapsed: 32.501 s <<< ERROR! java.io.FileNotFoundException: File does 
> not exist: 
> hdfs://localhost:46072/user/agent04_azpcontainer/.flink/application_1606950278664_0001/flink-dist_2.11-1.12-SNAPSHOT.jar
>  at 
> org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1441)
>  at 
> org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1434)
>  at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>  at 
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1434)
>  at 
> org.apache.flink.yarn.YARNFileReplicationITCase.extraVerification(YARNFileReplicationITCase.java:148)
>  at 
> org.apache.flink.yarn.YARNFileReplicationITCase.deployPerJob(YARNFileReplicationITCase.java:113)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong merged pull request #13800: [FLINK-19650][connectors jdbc]Support the limit push down for the Jdb…

2020-12-26 Thread GitBox


wuchong merged pull request #13800:
URL: https://github.com/apache/flink/pull/13800


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-19650) Support the limit push down for the Jdbc connector

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-19650.
---
Fix Version/s: 1.13.0
   Resolution: Fixed

Fixed in master: a859e2cca7961cc966cf0a3990e821a62fb52654

> Support the limit push down for the Jdbc connector
> --
>
> Key: FLINK-19650
> URL: https://issues.apache.org/jira/browse/FLINK-19650
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC, Table SQL / API
>Reporter: Shengkai Fang
>Assignee: Shengkai Fang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently the blink planner has already supported rule 
> {{PushLimitIntoLegacyTableSourceScanRule}}. It's ready to add this feature 
> for the jdbc connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong merged pull request #14466: [FLINK-20657][connectors/jdbc] Migrate jdbc InputFormat/LookupFunction to SimpleJdbcConnectionProvider for connection establishment

2020-12-26 Thread GitBox


wuchong merged pull request #14466:
URL: https://github.com/apache/flink/pull/14466


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-20657) Migrate jdbc connection code of JdbcInputFormat, JdbcLookupFunction, JdbcRowDataLookupFunction to SimpleJdbcConnectionProvider

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20657?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-20657.
---
Resolution: Fixed

Fixed in master: 0c36b666094c01a96d341d2b0ee0b6863a6484e8


> Migrate jdbc connection code of JdbcInputFormat, JdbcLookupFunction, 
> JdbcRowDataLookupFunction to SimpleJdbcConnectionProvider
> --
>
> Key: FLINK-20657
> URL: https://issues.apache.org/jira/browse/FLINK-20657
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / JDBC, Table SQL / Ecosystem
>Affects Versions: 1.13.0
>Reporter: Kezhu Wang
>Assignee: Kezhu Wang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, there are multiple places to establish jdbc connection, several of 
> them were written before introducing of {{JdbcConnectionProvider}} in 
> FLINK-15782. We need to migrate these classes for reasons:
> * Code reusability.
> * Circumvent pitfalls of {{DriverManager}} such as FLINK-19435 and possible 
> [others|https://ci.apache.org/projects/tomcat/tomcat85/docs/jndi-datasource-examples-howto.html#DriverManager,_the_service_provider_mechanism_and_memory_leaks]
>  in one place.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-20773) Support allow-unescaped-control-chars option for JSON format

2020-12-26 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17255172#comment-17255172
 ] 

Jark Wu commented on FLINK-20773:
-

Sounds good to me, as Jackson has also this feature 
{{JsonReadFeature#ALLOW_UNESCAPED_CONTROL_CHARS}}.

I assigned this issue to you [~xiaozilong]. 

> Support allow-unescaped-control-chars option for JSON format
> 
>
> Key: FLINK-20773
> URL: https://issues.apache.org/jira/browse/FLINK-20773
> Project: Flink
>  Issue Type: Improvement
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0
>Reporter: xiaozilong
>Priority: Major
> Attachments: image-2020-12-25-20-21-50-637.png
>
>
> Can we add an option `allow-unescaped-control-chars` for json format because 
> of it will throw exception when exist illegal unquoted characters in the data.
> !image-2020-12-25-20-21-50-637.png!
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20773) Support allow-unescaped-control-chars option for JSON format

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20773?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-20773:

Fix Version/s: 1.13.0

> Support allow-unescaped-control-chars option for JSON format
> 
>
> Key: FLINK-20773
> URL: https://issues.apache.org/jira/browse/FLINK-20773
> Project: Flink
>  Issue Type: Improvement
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0
>Reporter: xiaozilong
>Assignee: xiaozilong
>Priority: Major
> Fix For: 1.13.0
>
> Attachments: image-2020-12-25-20-21-50-637.png
>
>
> Can we add an option `allow-unescaped-control-chars` for json format because 
> of it will throw exception when exist illegal unquoted characters in the data.
> !image-2020-12-25-20-21-50-637.png!
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-20773) Support allow-unescaped-control-chars option for JSON format

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20773?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-20773:
---

Assignee: xiaozilong

> Support allow-unescaped-control-chars option for JSON format
> 
>
> Key: FLINK-20773
> URL: https://issues.apache.org/jira/browse/FLINK-20773
> Project: Flink
>  Issue Type: Improvement
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0
>Reporter: xiaozilong
>Assignee: xiaozilong
>Priority: Major
> Attachments: image-2020-12-25-20-21-50-637.png
>
>
> Can we add an option `allow-unescaped-control-chars` for json format because 
> of it will throw exception when exist illegal unquoted characters in the data.
> !image-2020-12-25-20-21-50-637.png!
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16175) Add config option to switch case sensitive for column names in SQL

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-16175:

Fix Version/s: (was: 1.13.0)

> Add config option to switch case sensitive for column names in SQL
> --
>
> Key: FLINK-16175
> URL: https://issues.apache.org/jira/browse/FLINK-16175
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API, Table SQL / Planner
>Affects Versions: 1.11.0
>Reporter: Leonard Xu
>Assignee: Leonard Xu
>Priority: Major
>  Labels: pull-request-available, usability
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Flink SQL is default CaseSensitive and have no option to config. This issue 
> aims to support
> a configOption so that user can set CaseSensitive for their SQL.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-17099) Refactoring State TTL solution in Group Agg、Deduplication、TopN operators replace Timer with StateTtlConfig

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17099?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-17099.
---
Fix Version/s: (was: 1.13.0)
   Resolution: Duplicate

This has been fixed by FLINK-17096. But note that TopN is not migrated to the 
StateTtlConfig becuase it holds multiple states and needs to sync with each 
other. 

> Refactoring State TTL solution in Group Agg、Deduplication、TopN operators 
> replace Timer with StateTtlConfig
> --
>
> Key: FLINK-17099
> URL: https://issues.apache.org/jira/browse/FLINK-17099
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Runtime
>Affects Versions: 1.9.0, 1.10.0
>Reporter: dalongliu
>Assignee: dalongliu
>Priority: Major
>
> At the moment, there are 2 ways to cleanup states.
> 1) registering a processing-time timer, and cleanup entries when the timer is 
> callback.
>  - pros: can cleanup multiple states at the same time (state consistent)
>  - cons: timer space depends on the key size, which may lead to OOM (heap 
> timer).
>  - used in Group Aggregation, Over Aggregation, TopN
> 2) using the {{StateTtlConfig}} provided by DataStream [1].
>  - pros: decouple the logic of state ttl with the record processing, easy to 
> program (take a look at old planner NonWindowJoin which bundles ttl timestamp 
> with records in MapState).
>  - cons: can't cleanup multiple states at the same time.
>  - useed in Sream-Stream Joins.
> For timer solution, although it can cleanup multiple states at the same time, 
> but it also will lead to OOM when there have a great many state keys, 
> besides, StateTtlConfig is used in stream-stream join case, and will be used 
> in more operator. Therefore,in order to unify the state ttl solution, 
> simplify the code implemention, and improve the readability of codes, so we 
> should refactor state cleanup way which use StateTtlConfig to replace 
> processing-time timer in Group Aggregation、Deduplication、TopN operators.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15064) Remove XmlOutput util class in blink planner

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-15064:

Fix Version/s: (was: 1.13.0)

> Remove XmlOutput util class in blink planner
> 
>
> Key: FLINK-15064
> URL: https://issues.apache.org/jira/browse/FLINK-15064
> Project: Flink
>  Issue Type: Task
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Priority: Major
>
> {{org.apache.flink.table.planner.utils.XmlOutput}} is introduced in blink 
> planner to fix a redundant line break for each CDATA section. This has been 
> fixed in Calcite via 1745f752561be04ae34d1fa08593c2d3ba4470e8. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-9477) FLIP-90: Support SQL 2016 JSON functions in Flink SQL

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-9477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-9477:
---
Flagged: Impediment

> FLIP-90: Support SQL 2016 JSON functions in Flink SQL
> -
>
> Key: FLINK-9477
> URL: https://issues.apache.org/jira/browse/FLINK-9477
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Shuyi Chen
>Assignee: Forward Xu
>Priority: Major
> Fix For: 1.13.0
>
>
> FLIP Link 
> https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=141724550



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-9477) FLIP-90: Support SQL 2016 JSON functions in Flink SQL

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-9477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-9477:
---
Flagged:   (was: Impediment)

> FLIP-90: Support SQL 2016 JSON functions in Flink SQL
> -
>
> Key: FLINK-9477
> URL: https://issues.apache.org/jira/browse/FLINK-9477
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Shuyi Chen
>Assignee: Forward Xu
>Priority: Major
> Fix For: 1.13.0
>
>
> FLIP Link 
> https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=141724550



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18778) Support the SupportsProjectionPushDown interface for LookupTableSource

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18778:

Fix Version/s: 1.13.0

> Support the SupportsProjectionPushDown interface for LookupTableSource
> --
>
> Key: FLINK-18778
> URL: https://issues.apache.org/jira/browse/FLINK-18778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Assignee: hailong wang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20774) Test

2020-12-26 Thread Jark Wu (Jira)
Jark Wu created FLINK-20774:
---

 Summary: Test
 Key: FLINK-20774
 URL: https://issues.apache.org/jira/browse/FLINK-20774
 Project: Flink
  Issue Type: New Feature
Reporter: Jark Wu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-20774) Test

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-20774.
---
Resolution: Invalid

> Test
> 
>
> Key: FLINK-20774
> URL: https://issues.apache.org/jira/browse/FLINK-20774
> Project: Flink
>  Issue Type: New Feature
>Reporter: Jark Wu
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16201) Support JSON_VALUE for blink planner

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-16201:

Labels: pull-request-available sprint  (was: pull-request-available)

> Support JSON_VALUE for blink planner
> 
>
> Key: FLINK-16201
> URL: https://issues.apache.org/jira/browse/FLINK-16201
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Zili Chen
>Assignee: Forward Xu
>Priority: Major
>  Labels: pull-request-available, sprint
> Fix For: 1.13.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20562) Support ExplainDetails for EXPLAIN sytnax

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-20562:

Labels: sprint  (was: )

> Support ExplainDetails for EXPLAIN sytnax
> -
>
> Key: FLINK-20562
> URL: https://issues.apache.org/jira/browse/FLINK-20562
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Jark Wu
>Priority: Major
>  Labels: sprint
> Fix For: 1.13.0
>
>
> Currently, {{EXPLAIN}} syntax only supports to print the default AST, logical 
> plan, and physical plan. However, it doesn't support to print detailed 
> information such as CHANGELOG_MODE, ESTIMATED_COST, JSON_EXECUTION_PLAN which 
> are defined in {{ExplainDetail}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20348) Make "schema-registry.subject" optional for Kafka sink with avro-confluent format

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-20348:

Labels: sprint  (was: )

> Make "schema-registry.subject" optional for Kafka sink with avro-confluent 
> format
> -
>
> Key: FLINK-20348
> URL: https://issues.apache.org/jira/browse/FLINK-20348
> Project: Flink
>  Issue Type: Improvement
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Ecosystem
>Reporter: Jark Wu
>Priority: Major
>  Labels: sprint
> Fix For: 1.13.0
>
>
> Currently, configuration "schema-registry.subject" in avro-confluent format 
> is required by sink. However, this is quite verbose set it manually. By 
> default, it can be to set to {{-key}} and {{-value}} 
> if it works with kafka or upsert-kafka connector. This can also makes 
> 'avro-confluent' format to be more handy and works better with 
> Kafka/Confluent ecosystem. 
> {code:sql}
> CREATE TABLE kafka_gmv (
>   day_str STRING,
>   gmv BIGINT,
>   PRIMARY KEY (day_str) NOT ENFORCED
> ) WITH (
> 'connector' = 'upsert-kafka',
> 'topic' = 'kafka_gmv',
> 'properties.bootstrap.servers' = 'localhost:9092',
> -- 'key.format' = 'raw',
> 'key.format' = 'avro-confluent',
> 'key.avro-confluent.schema-registry.url' = 'http://localhost:8181',
> 'key.avro-confluent.schema-registry.subject' = 'kafka_gmv-key',
> 'value.format' = 'avro-confluent',
> 'value.avro-confluent.schema-registry.url' = 'http://localhost:8181',
> 'value.avro-confluent.schema-registry.subject' = 'kafka_gmv-value'
> );
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17855) UDF with parameter Array(Row) can not work

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17855?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-17855:

Labels: sprint  (was: )

> UDF with parameter Array(Row) can not work
> --
>
> Key: FLINK-17855
> URL: https://issues.apache.org/jira/browse/FLINK-17855
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Jingsong Lee
>Priority: Major
>  Labels: sprint
> Fix For: 1.13.0
>
>
> {code:java}
> public String eval(Row[] rows) {
>   ...
> }
> {code}
> Can not work.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14364) Allow comments fail when not ignore parse errors in CsvRowDeserializationSchema

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-14364:

Fix Version/s: (was: 1.13.0)

> Allow comments fail when not ignore parse errors in 
> CsvRowDeserializationSchema
> ---
>
> Key: FLINK-14364
> URL: https://issues.apache.org/jira/browse/FLINK-14364
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Jingsong Lee
>Assignee: Jiayi Liao
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Use CsvRowDeserializationSchema, when setIgnoreParseErrors(false) and 
> setAllowComments(true).
> If there are some comments in msg, will throw MismatchedInputException.
> If this a bug? and we should catch MismatchedInputException and return null?
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14364) Allow comments fail when not ignore parse errors in CsvRowDeserializationSchema

2020-12-26 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14364?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17255178#comment-17255178
 ] 

Jark Wu commented on FLINK-14364:
-

According to {{CsvRowDataSerDeSchemaTest#testDeserializeIgnoreComment}} and 
{{CsvRowDataSerDeSchemaTest#testDeserializeAllowComment}}, I think this has 
been fixed in the new CSV format? Could we close this?

> Allow comments fail when not ignore parse errors in 
> CsvRowDeserializationSchema
> ---
>
> Key: FLINK-14364
> URL: https://issues.apache.org/jira/browse/FLINK-14364
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Jingsong Lee
>Assignee: Jiayi Liao
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Use CsvRowDeserializationSchema, when setIgnoreParseErrors(false) and 
> setAllowComments(true).
> If there are some comments in msg, will throw MismatchedInputException.
> If this a bug? and we should catch MismatchedInputException and return null?
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-20321) Get NPE when using AvroDeserializationSchema to deserialize null input

2020-12-26 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17255179#comment-17255179
 ] 

Jark Wu commented on FLINK-20321:
-

Hi [~sampadsaha5], are you still working on this? What's the status of this 
issue?

> Get NPE when using AvroDeserializationSchema to deserialize null input
> --
>
> Key: FLINK-20321
> URL: https://issues.apache.org/jira/browse/FLINK-20321
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Ecosystem
>Affects Versions: 1.12.0
>Reporter: Shengkai Fang
>Assignee: Sampad Kumar Saha
>Priority: Major
>  Labels: starter
> Fix For: 1.13.0
>
>
> You can reproduce the bug by adding the code into the 
> {{AvroDeserializationSchemaTest}}.
> The code follows
> {code:java}
> @Test
>   public void testSpecificRecord2() throws Exception {
>   DeserializationSchema deserializer = 
> AvroDeserializationSchema.forSpecific(Address.class);
>   Address deserializedAddress = deserializer.deserialize(null);
>   assertEquals(null, deserializedAddress);
>   }
> {code}
> Exception stack:
> {code:java}
> java.lang.NullPointerException
>   at 
> org.apache.flink.formats.avro.utils.MutableByteArrayInputStream.setBuffer(MutableByteArrayInputStream.java:43)
>   at 
> org.apache.flink.formats.avro.AvroDeserializationSchema.deserialize(AvroDeserializationSchema.java:131)
>   at 
> org.apache.flink.formats.avro.AvroDeserializationSchemaTest.testSpecificRecord2(AvroDeserializationSchemaTest.java:69)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20321) Get NPE when using AvroDeserializationSchema to deserialize null input

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-20321:

Labels: sprint starter  (was: starter)

> Get NPE when using AvroDeserializationSchema to deserialize null input
> --
>
> Key: FLINK-20321
> URL: https://issues.apache.org/jira/browse/FLINK-20321
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Ecosystem
>Affects Versions: 1.12.0
>Reporter: Shengkai Fang
>Assignee: Sampad Kumar Saha
>Priority: Major
>  Labels: sprint, starter
> Fix For: 1.13.0
>
>
> You can reproduce the bug by adding the code into the 
> {{AvroDeserializationSchemaTest}}.
> The code follows
> {code:java}
> @Test
>   public void testSpecificRecord2() throws Exception {
>   DeserializationSchema deserializer = 
> AvroDeserializationSchema.forSpecific(Address.class);
>   Address deserializedAddress = deserializer.deserialize(null);
>   assertEquals(null, deserializedAddress);
>   }
> {code}
> Exception stack:
> {code:java}
> java.lang.NullPointerException
>   at 
> org.apache.flink.formats.avro.utils.MutableByteArrayInputStream.setBuffer(MutableByteArrayInputStream.java:43)
>   at 
> org.apache.flink.formats.avro.AvroDeserializationSchema.deserialize(AvroDeserializationSchema.java:131)
>   at 
> org.apache.flink.formats.avro.AvroDeserializationSchemaTest.testSpecificRecord2(AvroDeserializationSchemaTest.java:69)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17868) Proctime in DDL can not work in batch mode

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17868?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-17868:

Labels: sprint  (was: )

> Proctime in DDL can not work in batch mode
> --
>
> Key: FLINK-17868
> URL: https://issues.apache.org/jira/browse/FLINK-17868
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Jingsong Lee
>Priority: Major
>  Labels: sprint
> Fix For: 1.13.0
>
>
> The data of this proctime column will be all null. Should same to current 
> timestamp.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19062) Improve alter table DDL

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19062?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-19062:

Labels: starter  (was: )

> Improve alter table DDL
> ---
>
> Key: FLINK-19062
> URL: https://issues.apache.org/jira/browse/FLINK-19062
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Jingsong Lee
>Priority: Major
>  Labels: starter
> Fix For: 1.13.0
>
>
> Support:
>  * ALTER TABLE table_name UNSET PROPERTIES (key)
>  * ALTER TABLE table_name CHANGE [COLUMN] col_old_name col_new_name 
> column_type [COMMENT col_comment]
>  * ALTER TABLE table_name add `column` ...
>  * ALTER TABLE table_name replace `column` ...



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19062) Improve alter table DDL

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19062?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-19062:

Labels:   (was: starter)

> Improve alter table DDL
> ---
>
> Key: FLINK-19062
> URL: https://issues.apache.org/jira/browse/FLINK-19062
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Jingsong Lee
>Priority: Major
> Fix For: 1.13.0
>
>
> Support:
>  * ALTER TABLE table_name UNSET PROPERTIES (key)
>  * ALTER TABLE table_name CHANGE [COLUMN] col_old_name col_new_name 
> column_type [COMMENT col_comment]
>  * ALTER TABLE table_name add `column` ...
>  * ALTER TABLE table_name replace `column` ...



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16627) Support only generate non-null values when serializing into JSON

2020-12-26 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16627?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17255180#comment-17255180
 ] 

Jark Wu commented on FLINK-16627:
-

Hi [~jackray], are you still working on this issue? 
I agree with [~nilerzhou] that a new config option key style has been 
introduced in FLIP-122, we may need to come up with a new key for this option. 
{{'json.encode.ignore-null-fields'=false/true}} sounds good to me. As an 
alternative, {{'json.encode.include-null-fields'=true/false}} is also fine to 
me. 



> Support only generate non-null values when serializing into JSON
> 
>
> Key: FLINK-16627
> URL: https://issues.apache.org/jira/browse/FLINK-16627
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Planner
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Assignee: jackray wang
>Priority: Major
> Fix For: 1.13.0
>
>
> {code:java}
> //sql
> CREATE TABLE sink_kafka ( subtype STRING , svt STRING ) WITH (……)
> {code}
>  
> {code:java}
> //sql
> CREATE TABLE source_kafka ( subtype STRING , svt STRING ) WITH (……)
> {code}
>  
> {code:java}
> //scala udf
> class ScalaUpper extends ScalarFunction {
> def eval(str: String) : String= { 
>if(str == null){
>return ""
>}else{
>return str
>}
> }
> 
> }
> btenv.registerFunction("scala_upper", new ScalaUpper())
> {code}
>  
> {code:java}
> //sql
> insert into sink_kafka select subtype, scala_upper(svt)  from source_kafka
> {code}
>  
>  
> 
> Sometimes the svt's value is null, inert into kafkas json like  
> \{"subtype":"qin","svt":null}
> If the amount of data is small, it is acceptable,but we process 10TB of data 
> every day, and there may be many nulls in the json, which affects the 
> efficiency. If you can add a parameter to remove the null key when defining a 
> sinktable, the performance will be greatly improved
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16627) Support only generate non-null values when serializing into JSON

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-16627:

Labels: sprint  (was: )

> Support only generate non-null values when serializing into JSON
> 
>
> Key: FLINK-16627
> URL: https://issues.apache.org/jira/browse/FLINK-16627
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Planner
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Assignee: jackray wang
>Priority: Major
>  Labels: sprint
> Fix For: 1.13.0
>
>
> {code:java}
> //sql
> CREATE TABLE sink_kafka ( subtype STRING , svt STRING ) WITH (……)
> {code}
>  
> {code:java}
> //sql
> CREATE TABLE source_kafka ( subtype STRING , svt STRING ) WITH (……)
> {code}
>  
> {code:java}
> //scala udf
> class ScalaUpper extends ScalarFunction {
> def eval(str: String) : String= { 
>if(str == null){
>return ""
>}else{
>return str
>}
> }
> 
> }
> btenv.registerFunction("scala_upper", new ScalaUpper())
> {code}
>  
> {code:java}
> //sql
> insert into sink_kafka select subtype, scala_upper(svt)  from source_kafka
> {code}
>  
>  
> 
> Sometimes the svt's value is null, inert into kafkas json like  
> \{"subtype":"qin","svt":null}
> If the amount of data is small, it is acceptable,but we process 10TB of data 
> every day, and there may be many nulls in the json, which affects the 
> efficiency. If you can add a parameter to remove the null key when defining a 
> sinktable, the performance will be greatly improved
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20644) Check return type of ScalarFunction eval method shouldn't be void

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-20644:

Labels: sprint starter  (was: starter)

> Check return type of ScalarFunction eval method shouldn't be void
> -
>
> Key: FLINK-20644
> URL: https://issues.apache.org/jira/browse/FLINK-20644
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.11.1
> Environment: groupId:org.apache.flink
> artifactId:flink-table-api-scala-bridge_2.11
> version:1.11.1
>Reporter: shiyu
>Priority: Major
>  Labels: sprint, starter
> Fix For: 1.13.0
>
> Attachments: image-2020-12-17-16-04-15-131.png, 
> image-2020-12-17-16-07-39-827.png
>
>
> flink-table-api-scala-bridge_2.11
>   !image-2020-12-17-16-07-39-827.png!
> !image-2020-12-17-16-04-15-131.png!
> console:
> SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".SLF4J: Failed 
> to load class "org.slf4j.impl.StaticLoggerBinder".SLF4J: Defaulting to 
> no-operation (NOP) logger implementationSLF4J: See 
> http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.ERROR 
> StatusLogger Log4j2 could not find a logging implementation. Please add 
> log4j-core to the classpath. Using SimpleLogger to log to the console.../* 1 
> *//* 2 */      public class StreamExecCalc$13 extends 
> org.apache.flink.table.runtime.operators.AbstractProcessStreamOperator/* 3 */ 
>          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator \{/* 4 *//* 5 
> */        private final Object[] references;/* 6 */        private transient 
> org.apache.flink.table.runtime.typeutils.StringDataSerializer 
> typeSerializer$6;/* 7 */        private transient 
> org.apache.flink.table.data.util.DataFormatConverters.StringConverter 
> converter$9;/* 8 */        private transient 
> cn.bicon.tableapitest.udf.ScalarFunctionTest$HashCode 
> function_cn$bicon$tableapitest$udf$ScalarFunctionTest$HashCode$8999e79cc91b971a8777461fb7698c58;/*
>  9 */        private transient 
> org.apache.flink.table.data.util.DataFormatConverters.GenericConverter 
> converter$12;/* 10 */        org.apache.flink.table.data.BoxedWrapperRowData 
> out = new org.apache.flink.table.data.BoxedWrapperRowData(3);/* 11 */        
> private final org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);/* 12 *//* 
> 13 */        public StreamExecCalc$13(/* 14 */            Object[] 
> references,/* 15 */            
> org.apache.flink.streaming.runtime.tasks.StreamTask task,/* 16 */            
> org.apache.flink.streaming.api.graph.StreamConfig config,/* 17 */            
> org.apache.flink.streaming.api.operators.Output output,/* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {/* 19 */          this.references = 
> references;/* 20 */          typeSerializer$6 = 
> (((org.apache.flink.table.runtime.typeutils.StringDataSerializer) 
> references[0]));/* 21 */          converter$9 = 
> (((org.apache.flink.table.data.util.DataFormatConverters.StringConverter) 
> references[1]));/* 22 */          
> function_cn$bicon$tableapitest$udf$ScalarFunctionTest$HashCode$8999e79cc91b971a8777461fb7698c58
>  = (((cn.bicon.tableapitest.udf.ScalarFunctionTest$HashCode) 
> references[2]));/* 23 */          converter$12 = 
> (((org.apache.flink.table.data.util.DataFormatConverters.GenericConverter) 
> references[3]));/* 24 */          this.setup(task, config, output);/* 25 */   
>        if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {/* 26 */    
>         ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) 
> this)/* 27 */              
> .setProcessingTimeService(processingTimeService);/* 28 */          }/* 29 */  
>       }/* 30 *//* 31 */        @Override/* 32 */        public void open() 
> throws Exception \{/* 33 */          super.open();/* 34 */          /* 35 */  
>         
> function_cn$bicon$tableapitest$udf$ScalarFunctionTest$HashCode$8999e79cc91b971a8777461fb7698c58.open(new
>  org.apache.flink.table.functions.FunctionContext(getRuntimeContext()));/* 36 
> */                 /* 37 */        }/* 38 *//* 39 */        @Override/* 40 */ 
>        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception \{/* 41 */          
> org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();/* 42 */          /* 
> 43 */          org.apache.flink.table.data.binary.BinaryStringData field$5;/* 
> 44 */          boolean isNull$5;/* 45 */          
> org.apache.flink.table.d

[jira] [Closed] (FLINK-18345) Support filter push down for lookup table source

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18345?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-18345.
---
Fix Version/s: (was: 1.13.0)
   Resolution: Duplicate

> Support filter push down for lookup table source
> 
>
> Key: FLINK-18345
> URL: https://issues.apache.org/jira/browse/FLINK-18345
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Jingsong Lee
>Priority: Major
>
> Especially for Filesystem lookup table source, it stores all records in 
> memory, if there is filter push down, can reduce memory effectively for 
> lookup table source.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20644) Check return type of ScalarFunction eval method shouldn't be void

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-20644:

Labels: starter  (was: sprint starter)

> Check return type of ScalarFunction eval method shouldn't be void
> -
>
> Key: FLINK-20644
> URL: https://issues.apache.org/jira/browse/FLINK-20644
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.11.1
> Environment: groupId:org.apache.flink
> artifactId:flink-table-api-scala-bridge_2.11
> version:1.11.1
>Reporter: shiyu
>Priority: Major
>  Labels: starter
> Fix For: 1.13.0
>
> Attachments: image-2020-12-17-16-04-15-131.png, 
> image-2020-12-17-16-07-39-827.png
>
>
> flink-table-api-scala-bridge_2.11
>   !image-2020-12-17-16-07-39-827.png!
> !image-2020-12-17-16-04-15-131.png!
> console:
> SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".SLF4J: Failed 
> to load class "org.slf4j.impl.StaticLoggerBinder".SLF4J: Defaulting to 
> no-operation (NOP) logger implementationSLF4J: See 
> http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.ERROR 
> StatusLogger Log4j2 could not find a logging implementation. Please add 
> log4j-core to the classpath. Using SimpleLogger to log to the console.../* 1 
> *//* 2 */      public class StreamExecCalc$13 extends 
> org.apache.flink.table.runtime.operators.AbstractProcessStreamOperator/* 3 */ 
>          implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator \{/* 4 *//* 5 
> */        private final Object[] references;/* 6 */        private transient 
> org.apache.flink.table.runtime.typeutils.StringDataSerializer 
> typeSerializer$6;/* 7 */        private transient 
> org.apache.flink.table.data.util.DataFormatConverters.StringConverter 
> converter$9;/* 8 */        private transient 
> cn.bicon.tableapitest.udf.ScalarFunctionTest$HashCode 
> function_cn$bicon$tableapitest$udf$ScalarFunctionTest$HashCode$8999e79cc91b971a8777461fb7698c58;/*
>  9 */        private transient 
> org.apache.flink.table.data.util.DataFormatConverters.GenericConverter 
> converter$12;/* 10 */        org.apache.flink.table.data.BoxedWrapperRowData 
> out = new org.apache.flink.table.data.BoxedWrapperRowData(3);/* 11 */        
> private final org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);/* 12 *//* 
> 13 */        public StreamExecCalc$13(/* 14 */            Object[] 
> references,/* 15 */            
> org.apache.flink.streaming.runtime.tasks.StreamTask task,/* 16 */            
> org.apache.flink.streaming.api.graph.StreamConfig config,/* 17 */            
> org.apache.flink.streaming.api.operators.Output output,/* 18 */            
> org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {/* 19 */          this.references = 
> references;/* 20 */          typeSerializer$6 = 
> (((org.apache.flink.table.runtime.typeutils.StringDataSerializer) 
> references[0]));/* 21 */          converter$9 = 
> (((org.apache.flink.table.data.util.DataFormatConverters.StringConverter) 
> references[1]));/* 22 */          
> function_cn$bicon$tableapitest$udf$ScalarFunctionTest$HashCode$8999e79cc91b971a8777461fb7698c58
>  = (((cn.bicon.tableapitest.udf.ScalarFunctionTest$HashCode) 
> references[2]));/* 23 */          converter$12 = 
> (((org.apache.flink.table.data.util.DataFormatConverters.GenericConverter) 
> references[3]));/* 24 */          this.setup(task, config, output);/* 25 */   
>        if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {/* 26 */    
>         ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) 
> this)/* 27 */              
> .setProcessingTimeService(processingTimeService);/* 28 */          }/* 29 */  
>       }/* 30 *//* 31 */        @Override/* 32 */        public void open() 
> throws Exception \{/* 33 */          super.open();/* 34 */          /* 35 */  
>         
> function_cn$bicon$tableapitest$udf$ScalarFunctionTest$HashCode$8999e79cc91b971a8777461fb7698c58.open(new
>  org.apache.flink.table.functions.FunctionContext(getRuntimeContext()));/* 36 
> */                 /* 37 */        }/* 38 *//* 39 */        @Override/* 40 */ 
>        public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception \{/* 41 */          
> org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();/* 42 */          /* 
> 43 */          org.apache.flink.table.data.binary.BinaryStringData field$5;/* 
> 44 */          boolean isNull$5;/* 45 */          
> org.apache.flink.table.data.bina

[jira] [Updated] (FLINK-18779) Support the SupportsFilterPushDown interface for LookupTableSource

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18779:

Labels: pull-request-available  (was: pull-request-available sprint)

> Support the SupportsFilterPushDown interface for LookupTableSource
> --
>
> Key: FLINK-18779
> URL: https://issues.apache.org/jira/browse/FLINK-18779
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Assignee: Shengkai Fang
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18779) Support the SupportsFilterPushDown interface for LookupTableSource

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18779:

Labels: pull-request-available sprint  (was: pull-request-available)

> Support the SupportsFilterPushDown interface for LookupTableSource
> --
>
> Key: FLINK-18779
> URL: https://issues.apache.org/jira/browse/FLINK-18779
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Assignee: Shengkai Fang
>Priority: Major
>  Labels: pull-request-available, sprint
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18778) Support the SupportsProjectionPushDown interface for LookupTableSource

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18778:

Labels: pull-request-available sprint  (was: pull-request-available)

> Support the SupportsProjectionPushDown interface for LookupTableSource
> --
>
> Key: FLINK-18778
> URL: https://issues.apache.org/jira/browse/FLINK-18778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Assignee: hailong wang
>Priority: Major
>  Labels: pull-request-available, sprint
> Fix For: 1.13.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18202) Introduce Protobuf format

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18202?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18202:

Labels: pull-request-available sprint  (was: pull-request-available)

> Introduce Protobuf format
> -
>
> Key: FLINK-18202
> URL: https://issues.apache.org/jira/browse/FLINK-18202
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / API
>Reporter: Benchao Li
>Priority: Major
>  Labels: pull-request-available, sprint
> Fix For: 1.13.0
>
> Attachments: image-2020-06-15-17-18-03-182.png
>
>
> PB[1] is a very famous and wildly used (de)serialization framework. The ML[2] 
> also has some discussions about this. It's a useful feature.
> This issue maybe needs some designs, or a FLIP.
> [1] [https://developers.google.com/protocol-buffers]
> [2] [http://apache-flink.147419.n8.nabble.com/Flink-SQL-UDF-td3725.html]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18202) Introduce Protobuf format

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18202?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18202:

Fix Version/s: 1.13.0

> Introduce Protobuf format
> -
>
> Key: FLINK-18202
> URL: https://issues.apache.org/jira/browse/FLINK-18202
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / API
>Reporter: Benchao Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
> Attachments: image-2020-06-15-17-18-03-182.png
>
>
> PB[1] is a very famous and wildly used (de)serialization framework. The ML[2] 
> also has some discussions about this. It's a useful feature.
> This issue maybe needs some designs, or a FLIP.
> [1] [https://developers.google.com/protocol-buffers]
> [2] [http://apache-flink.147419.n8.nabble.com/Flink-SQL-UDF-td3725.html]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-18202) Introduce Protobuf format

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18202?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-18202:
---

Assignee: Suhan Mao

> Introduce Protobuf format
> -
>
> Key: FLINK-18202
> URL: https://issues.apache.org/jira/browse/FLINK-18202
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / API
>Reporter: Benchao Li
>Assignee: Suhan Mao
>Priority: Major
>  Labels: pull-request-available, sprint
> Fix For: 1.13.0
>
> Attachments: image-2020-06-15-17-18-03-182.png
>
>
> PB[1] is a very famous and wildly used (de)serialization framework. The ML[2] 
> also has some discussions about this. It's a useful feature.
> This issue maybe needs some designs, or a FLIP.
> [1] [https://developers.google.com/protocol-buffers]
> [2] [http://apache-flink.147419.n8.nabble.com/Flink-SQL-UDF-td3725.html]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-16201) Support JSON_VALUE for blink planner

2020-12-26 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-16201:
---

Assignee: (was: Forward Xu)

> Support JSON_VALUE for blink planner
> 
>
> Key: FLINK-16201
> URL: https://issues.apache.org/jira/browse/FLINK-16201
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Zili Chen
>Priority: Major
>  Labels: pull-request-available, sprint
> Fix For: 1.13.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #8952: [FLINK-10868][flink-runtime] Add failure rater for resource manager

2020-12-26 Thread GitBox


flinkbot edited a comment on pull request #8952:
URL: https://github.com/apache/flink/pull/8952#issuecomment-513724324


   
   ## CI report:
   
   * d083b630115604e34b0a74498890aedbff61b2a7 UNKNOWN
   * c6cd36a8dfd854ecf63b2d248a4a28f889c873ec Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=11341)
 
   * d58927642909f50571ed6242605aac564e074f89 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] fsk119 commented on a change in pull request #14437: [FLINK-20458][docs] translate gettingStarted.zh.md and correct spelling errors in gettingStarted.md

2020-12-26 Thread GitBox


fsk119 commented on a change in pull request #14437:
URL: https://github.com/apache/flink/pull/14437#discussion_r549073093



##
File path: docs/dev/table/sql/gettingStarted.zh.md
##
@@ -22,54 +22,54 @@ specific language governing permissions and limitations
 under the License.
 -->
 
-Flink SQL makes it simple to develop streaming applications using standard 
SQL. It is easy to learn Flink if you have ever worked with a database or SQL 
like system by remaining ANSI-SQL 2011 compliant. This tutorial will help you 
get started quickly with a Flink SQL development environment. 
+Flink SQL 使得使用标准 SQL 开发流应用程序变的简单。如果你曾经在工作中使用过兼容 ANSI-SQL 2011 的数据库或类似的 SQL 
系统,那么就很容易学习 Flink。本教程将帮助你快速入门 Flink SQL 开发环境。
  
 * This will be replaced by the TOC
 {:toc}
 
 
-### Prerequisetes 
+### 先决条件
 
-You only need to have basic knowledge of SQL to follow along. No other 
programming experience is assumed. 
+你只需要具备 SQL 的基础知识即可,不需要其他编程经验。
 
-### Installation
+### 安装
 
-There are multiple ways to install Flink. For experimentation, the most common 
option is to download the binaries and run them locally. You can follow the 
steps in [local installation]({%link try-flink/local_installation.zh.md %}) to 
set up an environment for the rest of the tutorial. 
+安装 Flink 有多种方式。为了实验,最常见的选择是下载二进制包并在本地运行。你可以按照[本地模式安装]({% link 
try-flink/local_installation.zh.md %})中的步骤为本教程的剩余部分设置环境。
 
-Once you're all set, use the following command to start a local cluster from 
the installation folder:
+完成所有设置后,在安装文件夹中使用以下命令启动本地集群:
 
 {% highlight bash %}
 ./bin/start-cluster.sh
 {% endhighlight %}
  
-Once started, the Flink WebUI on [localhost:8081](localhost:8081) is available 
locally, from which you can monitor the different jobs.
+启动完成后,就可以在本地访问 Flink WebUI [localhost:8081](localhost:8081),你可以通过它来监视不同的作业。

Review comment:
   读起来似乎不太通顺? 或者 另起一句:通过它,你可以监控不同的作业。

##
File path: docs/dev/table/sql/gettingStarted.zh.md
##
@@ -129,11 +128,11 @@ FROM employee_information
 GROUP BY dep_id;
  {% endhighlight %} 
 
-Such queries are considered _stateful_. Flink's advanced fault-tolerance 
mechanism will maintain internal state and consistency, so queries always 
return the correct result, even in the face of hardware failure. 
+这样的查询被认为是 _有状态的_。Flink 的高级容错机制将维持内部状态和一致性,因此即使遇到硬件故障,查询也始终返回正确结果。
 
-## Sink Tables
+## Sink 表
 
-When running this query, the SQL client provides output in real-time but in a 
read-only fashion. Storing results - to power a report or dashboard - requires 
writing out to another table. This can be achieved using an `INSERT INTO` 
statement. The table referenced in this clause is known as a sink table. An 
`INSERT INTO` statement will be submitted as a detached query to the Flink 
cluster. 
+当运行此查询时,SQL 客户端实时但是以只读方式提供输出。存储结果(为报表或仪表板提供数据来源)需要写到另一个表。这可以使用 `INSERT INTO` 
语句来实现。本节中引用的表称为 sink 表。`INSERT INTO` 语句将作为一个独立查询被提交到 Flink 集群中。

Review comment:
   "存储结果(为报表或仪表板提供数据来源)需要写到另一个表" -> "存储结果,作为报表或仪表板提供数据来源,需要写到另一个表" 

##
File path: docs/dev/table/sql/gettingStarted.zh.md
##
@@ -22,54 +22,54 @@ specific language governing permissions and limitations
 under the License.
 -->
 
-Flink SQL makes it simple to develop streaming applications using standard 
SQL. It is easy to learn Flink if you have ever worked with a database or SQL 
like system by remaining ANSI-SQL 2011 compliant. This tutorial will help you 
get started quickly with a Flink SQL development environment. 
+Flink SQL 使得使用标准 SQL 开发流应用程序变的简单。如果你曾经在工作中使用过兼容 ANSI-SQL 2011 的数据库或类似的 SQL 
系统,那么就很容易学习 Flink。本教程将帮助你快速入门 Flink SQL 开发环境。

Review comment:
   get started quickly with a Flink SQL development environment -> 在 Flink 
SQL 开发环境下快速入门

##
File path: docs/dev/table/sql/gettingStarted.zh.md
##
@@ -79,16 +79,15 @@ SELECT CURRENT_TIMESTAMP;
 
 {% top %}
 
-## Source Tables
+## Source 表
 
-As with all SQL engines, Flink queries operate on top of tables. 
-It differs from a traditional database because Flink does not manage data at 
rest locally; instead, its queries operate continuously over external tables. 
+与所有 SQL 引擎一样,Flink 查询在表上进行操作。与传统数据库不同,因为 Flink 不在本地管理静态数据;相反,它的查询在外部表上连续运行。
 
-Flink data processing pipelines begin with source tables. Source tables 
produce rows operated over during the query's execution; they are the tables 
referenced in the `FROM` clause of a query.  These could be Kafka topics, 
databases, filesystems, or any other system that Flink knows how to consume. 
+Flink 数据处理管道开始于 source 表。在查询执行期间,source 表产生操作的行;它们是查询时 `FROM` 子句中引用的表。这些表可能是 
Kafka 的 topics,数据库,文件系统,或 Flink 知道如何消费的任何其他系统。

Review comment:
   pipeline 在这里应该是流水线的意思。
   
   "Source tables produce rows operated over during the query's execution" -> 
Source 表产生在查询执行期间可以被操作的行 。
   
   “or any other system that Flink knows how to consume” -> 或者 任何其它 Flink 
知道如何消费的系统

##
File path: docs/dev/table/sql/gettingStarted.zh.md
##
@@ -113,13 +112,13 @@ SELECT * from employee_information WHERE DeptId = 1;
 
 {% top %}