[GitHub] [flink] flinkbot commented on pull request #19494: [FLINK-27267][contrib] Migrate tests to JUnit5

2022-04-15 Thread GitBox


flinkbot commented on PR #19494:
URL: https://github.com/apache/flink/pull/19494#issuecomment-1100499721

   
   ## CI report:
   
   * 4b20ff4a6cbbc48cc8ce3e1baa10befdee58840a UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-27267) [JUnit5 Migration] Module: flink-contrib

2022-04-15 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27267?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-27267:
---
Labels: pull-request-available  (was: )

> [JUnit5 Migration] Module: flink-contrib
> 
>
> Key: FLINK-27267
> URL: https://issues.apache.org/jira/browse/FLINK-27267
> Project: Flink
>  Issue Type: Sub-task
>  Components: Tests
>Affects Versions: 1.16.0
>Reporter: RocMarshal
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] RocMarshal opened a new pull request, #19494: [FLINK-27267][contrib] Migrate tests to JUnit5

2022-04-15 Thread GitBox


RocMarshal opened a new pull request, #19494:
URL: https://github.com/apache/flink/pull/19494

   
   
   ## What is the purpose of the change
   
   Migrate tests to JUnit5
   
   
   ## Brief change log
   
   Migrate tests to JUnit5
   
   
   ## Verifying this change
   
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / **no** / don't 
know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-27267) [JUnit5 Migration] Module: flink-contrib

2022-04-15 Thread RocMarshal (Jira)
RocMarshal created FLINK-27267:
--

 Summary: [JUnit5 Migration] Module: flink-contrib
 Key: FLINK-27267
 URL: https://issues.apache.org/jira/browse/FLINK-27267
 Project: Flink
  Issue Type: Sub-task
  Components: Tests
Affects Versions: 1.16.0
Reporter: RocMarshal






--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-21501) Sync Chinese documentation with FLINK-21343

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21501?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21501:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor auto-unassigned 
pull-request-available  (was: auto-deprioritized-major auto-unassigned 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Sync Chinese documentation with FLINK-21343
> ---
>
> Key: FLINK-21501
> URL: https://issues.apache.org/jira/browse/FLINK-21501
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.13.0
>Reporter: Dawid Wysakowicz
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned, pull-request-available
>
> We should update the Chinese documentation with changes introduced in 
> FLINK-21343



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-20561) Add documentation for `records-lag-max` metric.

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20561?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20561:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Add documentation for `records-lag-max` metric. 
> 
>
> Key: FLINK-20561
> URL: https://issues.apache.org/jira/browse/FLINK-20561
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 1.11.0, 1.12.0
>Reporter: xiaozilong
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> Currently, there are no metric description for kafka topic's lag in f[link 
> metrics 
> docs|https://ci.apache.org/projects/flink/flink-docs-release-1.11/monitoring/metrics.html#connectors].
>  But this metric was reported in flink actually. So we should add some docs 
> to guide the users to use it.
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-20778) [FLINK-20778] the comment for the type of kafka consumer is wrong at KafkaPartitionSplit

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20778:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor auto-unassigned 
pull-request-available  (was: auto-deprioritized-major auto-unassigned 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> [FLINK-20778] the comment for the type of kafka consumer  is wrong at 
> KafkaPartitionSplit
> -
>
> Key: FLINK-20778
> URL: https://issues.apache.org/jira/browse/FLINK-20778
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: Jeremy Mei
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned, pull-request-available
> Attachments: Screen Capture_select-area_20201228011944.png
>
>
> the current code:
> {code:java}
>   // Indicating the split should consume from the earliest.
>   public static final long LATEST_OFFSET = -1;
>   // Indicating the split should consume from the latest.
>   public static final long EARLIEST_OFFSET = -2;
> {code}
> should be adjusted as blew
> {code:java}
>   // Indicating the split should consume from the latest.
>   public static final long LATEST_OFFSET = -1;
>   // Indicating the split should consume from the earliest.
>   public static final long EARLIEST_OFFSET = -2;
> {code}



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-26104) KeyError: 'type_info' in PyFlink test

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26104?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-26104:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> KeyError: 'type_info' in PyFlink test
> -
>
> Key: FLINK-26104
> URL: https://issues.apache.org/jira/browse/FLINK-26104
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.14.3
>Reporter: Huang Xingbo
>Priority: Major
>  Labels: stale-major, test-stability
>
> {code:java}
> 2022-02-14T04:33:10.9891373Z Feb 14 04:33:10 E   Caused by: 
> java.lang.RuntimeException: Failed to create stage bundle factory! 
> INFO:root:Initializing Python harness: 
> /__w/1/s/flink-python/pyflink/fn_execution/beam/beam_boot.py --id=103-1 
> --provision_endpoint=localhost:46669
> 2022-02-14T04:33:10.9892470Z Feb 14 04:33:10 E   
> INFO:root:Starting up Python harness in a standalone process.
> 2022-02-14T04:33:10.9893079Z Feb 14 04:33:10 E   Traceback 
> (most recent call last):
> 2022-02-14T04:33:10.9894030Z Feb 14 04:33:10 E File 
> "/__w/1/s/flink-python/dev/.conda/lib/python3.7/runpy.py", line 193, in 
> _run_module_as_main
> 2022-02-14T04:33:10.9894791Z Feb 14 04:33:10 E   
> "__main__", mod_spec)
> 2022-02-14T04:33:10.9895653Z Feb 14 04:33:10 E File 
> "/__w/1/s/flink-python/dev/.conda/lib/python3.7/runpy.py", line 85, in 
> _run_code
> 2022-02-14T04:33:10.9896395Z Feb 14 04:33:10 E   
> exec(code, run_globals)
> 2022-02-14T04:33:10.9904913Z Feb 14 04:33:10 E File 
> "/__w/1/s/flink-python/pyflink/fn_execution/beam/beam_boot.py", line 116, in 
> 
> 2022-02-14T04:33:10.9930244Z Feb 14 04:33:10 E   from 
> pyflink.fn_execution.beam import beam_sdk_worker_main
> 2022-02-14T04:33:10.9931563Z Feb 14 04:33:10 E File 
> "/__w/1/s/flink-python/pyflink/fn_execution/beam/beam_sdk_worker_main.py", 
> line 21, in 
> 2022-02-14T04:33:10.9932630Z Feb 14 04:33:10 E   import 
> pyflink.fn_execution.beam.beam_operations # noqa # pylint: 
> disable=unused-import
> 2022-02-14T04:33:10.9933754Z Feb 14 04:33:10 E File 
> "/__w/1/s/flink-python/pyflink/fn_execution/beam/beam_operations.py", line 
> 23, in 
> 2022-02-14T04:33:10.9934415Z Feb 14 04:33:10 E   from 
> pyflink.fn_execution import flink_fn_execution_pb2
> 2022-02-14T04:33:10.9935335Z Feb 14 04:33:10 E File 
> "/__w/1/s/flink-python/pyflink/fn_execution/flink_fn_execution_pb2.py", line 
> 2581, in 
> 2022-02-14T04:33:10.9936378Z Feb 14 04:33:10 E   
> _SCHEMA_FIELDTYPE.fields_by_name['time_info'].containing_oneof = 
> _SCHEMA_FIELDTYPE.oneofs_by_name['type_info']
> 2022-02-14T04:33:10.9946519Z Feb 14 04:33:10 E   KeyError: 
> 'type_info'
> 2022-02-14T04:33:10.9947110Z Feb 14 04:33:10 E   
> 2022-02-14T04:33:10.9947911Z Feb 14 04:33:10 Eat 
> org.apache.flink.streaming.api.runners.python.beam.BeamPythonFunctionRunner.createStageBundleFactory(BeamPythonFunctionRunner.java:566)
> 2022-02-14T04:33:10.9949048Z Feb 14 04:33:10 Eat 
> org.apache.flink.streaming.api.runners.python.beam.BeamPythonFunctionRunner.open(BeamPythonFunctionRunner.java:255)
> 2022-02-14T04:33:10.9950162Z Feb 14 04:33:10 Eat 
> org.apache.flink.streaming.api.operators.python.AbstractPythonFunctionOperator.open(AbstractPythonFunctionOperator.java:131)
> 2022-02-14T04:33:10.9951344Z Feb 14 04:33:10 Eat 
> org.apache.flink.streaming.api.operators.python.AbstractOneInputPythonFunctionOperator.open(AbstractOneInputPythonFunctionOperator.java:116)
> 2022-02-14T04:33:10.9952487Z Feb 14 04:33:10 Eat 
> org.apache.flink.streaming.api.operators.python.PythonKeyedProcessOperator.open(PythonKeyedProcessOperator.java:121)
> 2022-02-14T04:33:10.9953561Z Feb 14 04:33:10 Eat 
> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:110)
> 2022-02-14T04:33:10.9954565Z Feb 14 04:33:10 Eat 
> 

[jira] [Updated] (FLINK-25442) HBaseConnectorITCase.testTableSink failed on azure

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25442?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-25442:
---
  Labels: auto-deprioritized-critical test-stability  (was: stale-critical 
test-stability)
Priority: Major  (was: Critical)

This issue was labeled "stale-critical" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Critical, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> HBaseConnectorITCase.testTableSink failed on azure
> --
>
> Key: FLINK-25442
> URL: https://issues.apache.org/jira/browse/FLINK-25442
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.15.0, 1.14.2, 1.16.0
>Reporter: Yun Gao
>Priority: Major
>  Labels: auto-deprioritized-critical, test-stability
>
> {code:java}
> Dec 24 00:48:54 Picked up JAVA_TOOL_OPTIONS: -XX:+HeapDumpOnOutOfMemoryError
> Dec 24 00:48:54 OpenJDK 64-Bit Server VM warning: ignoring option 
> MaxPermSize=128m; support was removed in 8.0
> Dec 24 00:48:54 Running org.apache.flink.connector.hbase2.HBaseConnectorITCase
> Dec 24 00:48:59 Formatting using clusterid: testClusterID
> Dec 24 00:50:15 java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10]
> Dec 24 00:50:15 Thread[HFileArchiver-8,5,PEWorkerGroup]
> Dec 24 00:50:15 Thread[HFileArchiver-9,5,PEWorkerGroup]
> Dec 24 00:50:15 Thread[HFileArchiver-10,5,PEWorkerGroup]
> Dec 24 00:50:15 Thread[HFileArchiver-11,5,PEWorkerGroup]
> Dec 24 00:50:15 Thread[HFileArchiver-12,5,PEWorkerGroup]
> Dec 24 00:50:15 Thread[HFileArchiver-13,5,PEWorkerGroup]
> Dec 24 00:50:16 Tests run: 9, Failures: 1, Errors: 0, Skipped: 0, Time 
> elapsed: 82.068 sec <<< FAILURE! - in 
> org.apache.flink.connector.hbase2.HBaseConnectorITCase
> Dec 24 00:50:16 
> testTableSink(org.apache.flink.connector.hbase2.HBaseConnectorITCase)  Time 
> elapsed: 8.534 sec  <<< FAILURE!
> Dec 24 00:50:16 java.lang.AssertionError: expected:<8> but was:<5>
> Dec 24 00:50:16   at org.junit.Assert.fail(Assert.java:89)
> Dec 24 00:50:16   at org.junit.Assert.failNotEquals(Assert.java:835)
> Dec 24 00:50:16   at org.junit.Assert.assertEquals(Assert.java:120)
> Dec 24 00:50:16   at org.junit.Assert.assertEquals(Assert.java:146)
> Dec 24 00:50:16   at 
> org.apache.flink.connector.hbase2.HBaseConnectorITCase.testTableSink(HBaseConnectorITCase.java:291)
> Dec 24 00:50:16   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Dec 24 00:50:16   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Dec 24 00:50:16   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Dec 24 00:50:16   at java.lang.reflect.Method.invoke(Method.java:498)
> Dec 24 00:50:16   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Dec 24 00:50:16   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Dec 24 00:50:16   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Dec 24 00:50:16   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Dec 24 00:50:16   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> Dec 24 00:50:16   at 
> org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> Dec 24 00:50:16   at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
> Dec 24 00:50:16   at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
> Dec 24 00:50:16   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
> Dec 24 00:50:16   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
> Dec 24 00:50:16   at 
> org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
> Dec 24 00:50:16   at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
> Dec 24 00:50:16   at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
> Dec 24 00:50:16   at 
> org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-20918) Avoid excessive flush of Hadoop output stream

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20918:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Avoid excessive flush of Hadoop output stream
> -
>
> Key: FLINK-20918
> URL: https://issues.apache.org/jira/browse/FLINK-20918
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hadoop Compatibility, FileSystems
>Affects Versions: 1.11.3, 1.12.0
>Reporter: Paul Lin
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> [HadoopRecoverableFsDataOutputStream#sync|https://github.com/apache/flink/blob/67d167ccd45046fc5ed222ac1f1e3ba5e6ec434b/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopRecoverableFsDataOutputStream.java#L123]
>  calls both `hflush` and `hsync`, whereas `hsync` is an enhanced version of 
> `hflush`. We should remove the `hflush` call to avoid the excessive flush.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-20938) Implement Flink's own tencent COS filesystem

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20938:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Implement Flink's own tencent COS filesystem
> 
>
> Key: FLINK-20938
> URL: https://issues.apache.org/jira/browse/FLINK-20938
> Project: Flink
>  Issue Type: New Feature
>  Components: FileSystems
>Affects Versions: 1.12.0
>Reporter: hayden zhou
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> Tencent's COS is widely used among China's cloud users, and Hadoop supports 
> Tencent COS since 3.3.0. 
> Open this jira to wrap CosNFileSystem in FLINK(similar to oss support), so 
> that users can read from & write to COS more easily in FLINK. 



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-26110) AvroStreamingFileSinkITCase failed on azure

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-26110:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> AvroStreamingFileSinkITCase failed on azure
> ---
>
> Key: FLINK-26110
> URL: https://issues.apache.org/jira/browse/FLINK-26110
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.13.5
>Reporter: Yun Gao
>Priority: Major
>  Labels: stale-major, test-stability
>
> {code:java}
> Feb 12 01:00:00 [ERROR] Tests run: 3, Failures: 1, Errors: 0, Skipped: 0, 
> Time elapsed: 2.433 s <<< FAILURE! - in 
> org.apache.flink.formats.avro.AvroStreamingFileSinkITCase
> Feb 12 01:00:00 [ERROR] 
> testWriteAvroGeneric(org.apache.flink.formats.avro.AvroStreamingFileSinkITCase)
>   Time elapsed: 0.433 s  <<< FAILURE!
> Feb 12 01:00:00 java.lang.AssertionError: expected:<1> but was:<2>
> Feb 12 01:00:00   at org.junit.Assert.fail(Assert.java:88)
> Feb 12 01:00:00   at org.junit.Assert.failNotEquals(Assert.java:834)
> Feb 12 01:00:00   at org.junit.Assert.assertEquals(Assert.java:645)
> Feb 12 01:00:00   at org.junit.Assert.assertEquals(Assert.java:631)
> Feb 12 01:00:00   at 
> org.apache.flink.formats.avro.AvroStreamingFileSinkITCase.validateResults(AvroStreamingFileSinkITCase.java:139)
> Feb 12 01:00:00   at 
> org.apache.flink.formats.avro.AvroStreamingFileSinkITCase.testWriteAvroGeneric(AvroStreamingFileSinkITCase.java:109)
> Feb 12 01:00:00   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Feb 12 01:00:00   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Feb 12 01:00:00   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Feb 12 01:00:00   at java.lang.reflect.Method.invoke(Method.java:498)
> Feb 12 01:00:00   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> Feb 12 01:00:00   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Feb 12 01:00:00   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> Feb 12 01:00:00   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Feb 12 01:00:00   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> Feb 12 01:00:00   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
> Feb 12 01:00:00   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
> Feb 12 01:00:00   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Feb 12 01:00:00   at java.lang.Thread.run(Thread.java:748)
> Feb 12 01:00:00 
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=31304=logs=c91190b6-40ae-57b2-5999-31b869b0a7c1=43529380-51b4-5e90-5af4-2dccec0ef402=13026



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-26047) Support usrlib in HDFS for YARN application mode

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-26047:
---
Labels: pull-request-available stale-assigned  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issue is assigned but has not 
received an update in 30 days, so it has been labeled "stale-assigned".
If you are still working on the issue, please remove the label and add a 
comment updating the community on your progress.  If this issue is waiting on 
feedback, please consider this a reminder to the committer/reviewer. Flink is a 
very active project, and so we appreciate your patience.
If you are no longer working on the issue, please unassign yourself so someone 
else may work on it.


> Support usrlib in HDFS for YARN application mode
> 
>
> Key: FLINK-26047
> URL: https://issues.apache.org/jira/browse/FLINK-26047
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Reporter: Biao Geng
>Assignee: Biao Geng
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> In YARN Application mode, we currently support using user jar and lib jar 
> from HDFS. For example, we can run commands like:
> {quote}./bin/flink run-application -t yarn-application \
>   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
>   hdfs://myhdfs/jars/my-application.jar{quote}
> For {{usrlib}}, we currently only support local directory. I propose to add 
> HDFS support for {{usrlib}} to work with CLASSPATH_INCLUDE_USER_JAR better. 
> It can also benefit cases like using notebook to submit flink job.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-21643) JDBC sink should be able to execute statements on multiple tables

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21643:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> JDBC sink should be able to execute statements on multiple tables
> -
>
> Key: FLINK-21643
> URL: https://issues.apache.org/jira/browse/FLINK-21643
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / JDBC
>Reporter: Maciej Obuchowski
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> Currently datastream JDBC sink supports outputting data only to one table - 
> by having to provide SQL template, from which SimpleBatchStatementExecutor 
> creates PreparedStatement. Creating multiple sinks, each of which writes data 
> to one table is impractical for moderate to large number of tables - 
> relational databases don't usually tolerate large number of connections.
> I propose adding DynamicBatchStatementExecutor, which will additionally 
> require
> 1) provided mechanism to create SQL statements based on given object
> 2) cache for prepared statements
> 3) mechanism for determining which statement should be used for given object



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-25529) java.lang.ClassNotFoundException: org.apache.orc.PhysicalWriter when write bulkly into hive-2.1.1 orc table

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-25529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-25529:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> java.lang.ClassNotFoundException: org.apache.orc.PhysicalWriter when write 
> bulkly into hive-2.1.1 orc table
> ---
>
> Key: FLINK-25529
> URL: https://issues.apache.org/jira/browse/FLINK-25529
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
> Environment: hive 2.1.1
> flink 1.12.4
>Reporter: Yuan Zhu
>Priority: Major
>  Labels: pull-request-available, stale-major
> Attachments: lib.jpg
>
>
> I tried to write data bulkly into hive-2.1.1 with orc format, and encountered 
> java.lang.ClassNotFoundException: org.apache.orc.PhysicalWriter
>  
> Using bulk writer by setting table.exec.hive.fallback-mapred-writer = false;
>  
> {code:java}
> SET 'table.sql-dialect'='hive';
> create table orders(
>     order_id int,
>     order_date timestamp,
>     customer_name string,
>     price decimal(10,3),
>     product_id int,
>     order_status boolean
> )partitioned by (dt string)
> stored as orc;
>  
> SET 'table.sql-dialect'='default';
> create table datagen_source (
> order_id int,
> order_date timestamp(9),
> customer_name varchar,
> price decimal(10,3),
> product_id int,
> order_status boolean
> )with('connector' = 'datagen');
> create catalog myhive with ('type' = 'hive', 'hive-conf-dir' = '/mnt/conf');
> set table.exec.hive.fallback-mapred-writer = false;
> insert into myhive.`default`.orders
> /*+ OPTIONS(
>     'sink.partition-commit.trigger'='process-time',
>     'sink.partition-commit.policy.kind'='metastore,success-file',
>     'sink.rolling-policy.file-size'='128MB',
>     'sink.rolling-policy.rollover-interval'='10s',
>     'sink.rolling-policy.check-interval'='10s',
>     'auto-compaction'='true',
>     'compaction.file-size'='1MB'    ) */
> select * , date_format(now(),'-MM-dd') as dt from datagen_source;  {code}
> [ERROR] Could not execute SQL statement. Reason:
> java.lang.ClassNotFoundException: org.apache.orc.PhysicalWriter
>  
> My jars in lib dir listed in attachment.
> In HiveTableSink#createStreamSink(line:270), createBulkWriterFactory if 
> table.exec.hive.fallback-mapred-writer is false.
> If table is orc, HiveShimV200#createOrcBulkWriterFactory will be invoked. 
> OrcBulkWriterFactory depends on org.apache.orc.PhysicalWriter in orc-core, 
> but flink-connector-hive excludes orc-core for conflicting with hive-exec.
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-20979) Wrong description of 'sink.bulk-flush.max-actions'

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20979?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-20979:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available starter  (was: auto-deprioritized-major 
pull-request-available stale-minor starter)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Wrong description of 'sink.bulk-flush.max-actions'
> --
>
> Key: FLINK-20979
> URL: https://issues.apache.org/jira/browse/FLINK-20979
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / ElasticSearch, Documentation, Table SQL / 
> Ecosystem
>Affects Versions: 1.12.0
>Reporter: Dawid Wysakowicz
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available, starter
>
> The documentation claims the option can be disabled with '0', but it can 
> actually be disable with '-1' whereas '0' is an illegal value.
> {{Maximum number of buffered actions per bulk request. Can be set to '0' to 
> disable it. }}



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-26109) Avro Confluent Schema Registry nightly end-to-end test failed on azure

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-26109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-26109:
---
Labels: stale-major test-stability  (was: test-stability)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Avro Confluent Schema Registry nightly end-to-end test failed on azure
> --
>
> Key: FLINK-26109
> URL: https://issues.apache.org/jira/browse/FLINK-26109
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.15.0
>Reporter: Yun Gao
>Priority: Major
>  Labels: stale-major, test-stability
>
> {code:java}
> Feb 12 07:55:02 Stopping job timeout watchdog (with pid=130662)
> Feb 12 07:55:03 Checking for errors...
> Feb 12 07:55:03 Found error in log files; printing first 500 lines; see full 
> logs for details:
> ...
> az209-567.vil1xujjdrkuxjp2ihtao45w0e.ax.internal.cloudapp.net 
> (dataPort=41161).
> org.apache.flink.util.FlinkException: The TaskExecutor is shutting down.
>     at 
> org.apache.flink.runtime.taskexecutor.TaskExecutor.onStop(TaskExecutor.java:456)
>  ~[flink-dist-1.15-SNAPSHOT.jar:1.15-SNAPSHOT]
>     at 
> org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStop(RpcEndpoint.java:214)
>  ~[flink-dist-1.15-SNAPSHOT.jar:1.15-SNAPSHOT]
>     at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StartedState.lambda$terminate$0(AkkaRpcActor.java:568)
>  ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at 
> org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:83)
>  ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StartedState.terminate(AkkaRpcActor.java:567)
>  ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleControlMessage(AkkaRpcActor.java:191)
>  ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:24) 
> ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:20) 
> ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at scala.PartialFunction.applyOrElse(PartialFunction.scala:123) 
> ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at scala.PartialFunction.applyOrElse$(PartialFunction.scala:122) 
> ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:20) 
> ~[flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.actor.Actor.aroundReceive(Actor.scala:537) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.actor.Actor.aroundReceive$(Actor.scala:535) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:220) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:580) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.actor.ActorCell.invoke(ActorCell.scala:548) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:270) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.dispatch.Mailbox.run(Mailbox.scala:231) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at akka.dispatch.Mailbox.exec(Mailbox.scala:243) 
> [flink-rpc-akka_7dcae025-2017-4b0f-828d-f89a7ceb9bf7.jar:1.15-SNAPSHOT]
>     at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) 
> [?:1.8.0_312]
>     at 
> java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056) 
> 

[jira] [Updated] (FLINK-21440) Translate Real Time Reporting with the Table API doc and correct a spelling mistake

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21440:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Translate Real Time Reporting with the Table API doc and correct a spelling 
> mistake
> ---
>
> Key: FLINK-21440
> URL: https://issues.apache.org/jira/browse/FLINK-21440
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation, Table SQL / Ecosystem
>Reporter: GuotaoLi
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> * Translate Real Time Reporting with the Table API doc to Chinese
> * Correct Real Time Reporting with the Table API doc allong with spelling 
> mistake to along with



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-21444) Lookup joins should deal with intermediate table scans correctly

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21444:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor auto-unassigned 
 (was: auto-deprioritized-major auto-unassigned stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Lookup joins should deal with intermediate table scans correctly
> 
>
> Key: FLINK-21444
> URL: https://issues.apache.org/jira/browse/FLINK-21444
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Caizhi Weng
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned
>
> Add the following test case to 
> {{org.apache.flink.table.planner.runtime.stream.sql.LookupJoinITCase}}
> {code:scala}
> @Test
> def myTest(): Unit = {
>   
> tEnv.getConfig.getConfiguration.setBoolean(OptimizerConfigOptions.TABLE_OPTIMIZER_REUSE_SOURCE_ENABLED,
>  true)
>   
> tEnv.getConfig.getConfiguration.setBoolean(RelNodeBlockPlanBuilder.TABLE_OPTIMIZER_REUSE_OPTIMIZE_BLOCK_WITH_DIGEST_ENABLED,
>  true)
>   val ddl1 =
> """
>   |CREATE TABLE sink1 (
>   |  `id` BIGINT
>   |) WITH (
>   |  'connector' = 'blackhole'
>   |)
>   |""".stripMargin
>   tEnv.executeSql(ddl1)
>   val ddl2 =
> """
>   |CREATE TABLE sink2 (
>   |  `id` BIGINT
>   |) WITH (
>   |  'connector' = 'blackhole'
>   |)
>   |""".stripMargin
>   tEnv.executeSql(ddl2)
>   val sql1 = "INSERT INTO sink1 SELECT T.id FROM src AS T JOIN user_table for 
> system_time as of T.proctime AS D ON T.id = D.id"
>   val sql2 = "INSERT INTO sink2 SELECT T.id FROM src AS T JOIN user_table for 
> system_time as of T.proctime AS D ON T.id + 1 = D.id"
>   val stmtSet = tEnv.createStatementSet()
>   stmtSet.addInsertSql(sql1)
>   stmtSet.addInsertSql(sql2)
>   stmtSet.execute().await()
> }
> {code}
> The following exception will occur
> {code}
> org.apache.flink.table.api.ValidationException: Temporal Table Join requires 
> primary key in versioned table, but no primary key can be found. The physical 
> plan is:
> FlinkLogicalJoin(condition=[AND(=($0, $2), 
> __INITIAL_TEMPORAL_JOIN_CONDITION($1, __TEMPORAL_JOIN_LEFT_KEY($0), 
> __TEMPORAL_JOIN_RIGHT_KEY($2)))], joinType=[inner])
>   FlinkLogicalCalc(select=[id, proctime])
> FlinkLogicalIntermediateTableScan(table=[[IntermediateRelTable_0]], 
> fields=[id, len, content, proctime])
>   FlinkLogicalSnapshot(period=[$cor0.proctime])
> FlinkLogicalCalc(select=[id])
>   FlinkLogicalIntermediateTableScan(table=[[IntermediateRelTable_1]], 
> fields=[age, id, name])
>   at 
> org.apache.flink.table.planner.plan.rules.logical.TemporalJoinRewriteWithUniqueKeyRule.org$apache$flink$table$planner$plan$rules$logical$TemporalJoinRewriteWithUniqueKeyRule$$validateRightPrimaryKey(TemporalJoinRewriteWithUniqueKeyRule.scala:124)
>   at 
> org.apache.flink.table.planner.plan.rules.logical.TemporalJoinRewriteWithUniqueKeyRule$$anon$1.visitCall(TemporalJoinRewriteWithUniqueKeyRule.scala:88)
>   at 
> org.apache.flink.table.planner.plan.rules.logical.TemporalJoinRewriteWithUniqueKeyRule$$anon$1.visitCall(TemporalJoinRewriteWithUniqueKeyRule.scala:70)
>   at org.apache.calcite.rex.RexCall.accept(RexCall.java:174)
>   at org.apache.calcite.rex.RexShuttle.visitList(RexShuttle.java:158)
>   at org.apache.calcite.rex.RexShuttle.visitCall(RexShuttle.java:110)
>   at 
> org.apache.flink.table.planner.plan.rules.logical.TemporalJoinRewriteWithUniqueKeyRule$$anon$1.visitCall(TemporalJoinRewriteWithUniqueKeyRule.scala:109)
>   at 
> org.apache.flink.table.planner.plan.rules.logical.TemporalJoinRewriteWithUniqueKeyRule$$anon$1.visitCall(TemporalJoinRewriteWithUniqueKeyRule.scala:70)
>   at org.apache.calcite.rex.RexCall.accept(RexCall.java:174)
>   at 
> org.apache.flink.table.planner.plan.rules.logical.TemporalJoinRewriteWithUniqueKeyRule.onMatch(TemporalJoinRewriteWithUniqueKeyRule.scala:70)
>   at 
> org.apache.calcite.plan.AbstractRelOptPlanner.fireRule(AbstractRelOptPlanner.java:333)
>   at org.apache.calcite.plan.hep.HepPlanner.applyRule(HepPlanner.java:542)
>   at 
> org.apache.calcite.plan.hep.HepPlanner.applyRules(HepPlanner.java:407)
>   at 
> org.apache.calcite.plan.hep.HepPlanner.executeInstruction(HepPlanner.java:243)
>   at 
> org.apache.calcite.plan.hep.HepInstruction$RuleInstance.execute(HepInstruction.java:127)
>   at 
> 

[jira] [Updated] (FLINK-21686) Duplicate code in hive parser file should be abstracted into functions

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21686?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21686:
---
  Labels: auto-deprioritized-minor pull-request-available  (was: 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Duplicate code in hive parser file should be abstracted into functions
> --
>
> Key: FLINK-21686
> URL: https://issues.apache.org/jira/browse/FLINK-21686
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive
>Reporter: humengyu
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, pull-request-available
>
> It would be better to use functions rather than duplicate code in hive parser 
> file:
>  #   option should be abstracted into a function;
>  #option should be abstracted into a function.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Commented] (FLINK-27255) Flink-avro does not support serialization and deserialization of avro schema longer than 65535 characters

2022-04-15 Thread Steven Zhen Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522943#comment-17522943
 ] 

Steven Zhen Wu commented on FLINK-27255:


[~jinyius] this issue existed for a while now. not sth new.

> Flink-avro does not support serialization and deserialization of avro schema 
> longer than 65535 characters
> -
>
> Key: FLINK-27255
> URL: https://issues.apache.org/jira/browse/FLINK-27255
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.14.4
>Reporter: Haizhou Zhao
>Assignee: Haizhou Zhao
>Priority: Major
>
> The underlying serialization of avro schema uses string serialization method 
> of ObjectOutputStream.class, however, the default string serialization by 
> ObjectOutputStream.class does not support handling string of more than 66535 
> characters (64kb). As a result, constructing flink operators that 
> input/output Avro Generic Record with huge schema is not possible.
>  
> The purposed fix is two change the serialization and deserialization method 
> of these following classes so that huge string could also be handled.
>  
> [GenericRecordAvroTypeInfo|https://github.com/apache/flink/blob/master/flink-formats/flink-avro/src/main/java/org/apache/flink/formats/avro/typeutils/GenericRecordAvroTypeInfo.java#L107]
> [SerializableAvroSchema|https://github.com/apache/flink/blob/master/flink-formats/flink-avro/src/main/java/org/apache/flink/formats/avro/typeutils/SerializableAvroSchema.java#L55]
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink-kubernetes-operator] gyfora commented on pull request #165: [FLINK-26140] Support rollback strategies

2022-04-15 Thread GitBox


gyfora commented on PR #165:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/165#issuecomment-1100367932

   @wangyang0918 I think for the suspend to work, I need to fix 2 things 
basically:
1. If the user suspends a job, the spec should be automatically marked as 
`lastStableSpec`
2. If an upgrade after a suspend fails, we should roll back to the 
suspended state (simply stop the job)

Will work on this over the weekend


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] metaswirl commented on a diff in pull request #17228: [FLINK-24236] Migrate tests to factory approach

2022-04-15 Thread GitBox


metaswirl commented on code in PR #17228:
URL: https://github.com/apache/flink/pull/17228#discussion_r851495599


##
flink-runtime/src/test/java/org/apache/flink/runtime/metrics/ReporterSetupTest.java:
##
@@ -47,17 +50,30 @@
 import static org.junit.Assert.assertTrue;
 
 /** Tests for the {@link ReporterSetup}. */
-public class ReporterSetupTest extends TestLogger {
+@ExtendWith(TestLoggerExtension.class)
+class ReporterSetupTest {
+
+@RegisterExtension
+static final ContextClassLoaderExtension CONTEXT_CLASS_LOADER_EXTENSION =
+ContextClassLoaderExtension.builder()
+.withServiceEntry(
+MetricReporterFactory.class,
+TestReporter1.class.getName(),
+TestReporter2.class.getName(),
+TestReporter11.class.getName(),
+TestReporter12.class.getName(),
+TestReporter13.class.getName())
+.build();

Review Comment:
   What is the benefit of using a custom class loader here?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #19493: [FLINK-27205][docs-zh] Translate "Concepts -> Glossary" page into Chinese.

2022-04-15 Thread GitBox


flinkbot commented on PR #19493:
URL: https://github.com/apache/flink/pull/19493#issuecomment-1100198258

   
   ## CI report:
   
   * 42bb3147bb7ab02a51fdc0c0b9cca28c15289a91 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] liuzhuang2017 opened a new pull request, #19493: [FLINK-27205][docs-zh] Translate "Concepts -> Glossary" page into Chinese.

2022-04-15 Thread GitBox


liuzhuang2017 opened a new pull request, #19493:
URL: https://github.com/apache/flink/pull/19493

   
   ## What is the purpose of the change
   
   
![image](https://user-images.githubusercontent.com/95120044/163592412-7e4e94b7-1800-422a-8675-e78b34c0601d.png)
   
   
![image](https://user-images.githubusercontent.com/95120044/163592580-51582550-1f40-4549-a56f-43b6ec1c3475.png)
   
   
   - **As can be seen from the above figure, there are still some untranslated 
content in the Chinese document, so this part is translated.**
   
   
   ## Brief change log
   
   Translate "Concepts -> Glossary" page into Chinese.
   
   ## Verifying this change
   This change is a trivial rework / code cleanup without any test coverage.
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-27264) Add ITCase for concurrent batch overwrite and streaming insert

2022-04-15 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-27264:
---
Labels: pull-request-available  (was: )

> Add ITCase for concurrent batch overwrite and streaming insert
> --
>
> Key: FLINK-27264
> URL: https://issues.apache.org/jira/browse/FLINK-27264
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table Store
>Affects Versions: table-store-0.1.0
>Reporter: Jane Chan
>Assignee: Jane Chan
>Priority: Major
>  Labels: pull-request-available
> Fix For: table-store-0.1.0
>
> Attachments: image-2022-04-15-19-26-09-649.png
>
>
> !image-2022-04-15-19-26-09-649.png|width=609,height=241!
> Add it case for user story



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink-table-store] LadyForest opened a new pull request, #91: [FLINK-27264] Add IT case for concurrent batch overwrite and streaming insert into

2022-04-15 Thread GitBox


LadyForest opened a new pull request, #91:
URL: https://github.com/apache/flink-table-store/pull/91

   Add it case to cover the user story


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-web] liuzhuang2017 commented on pull request #521: [hotfix][docs] Fix link tags typo.

2022-04-15 Thread GitBox


liuzhuang2017 commented on PR #521:
URL: https://github.com/apache/flink-web/pull/521#issuecomment-1100116349

   @wuchong  ,Hi,please help me review this pr when you are free time, thank 
you.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #19492: Update TimeWindow.java

2022-04-15 Thread GitBox


flinkbot commented on PR #19492:
URL: https://github.com/apache/flink/pull/19492#issuecomment-1100089900

   
   ## CI report:
   
   * e5b4f082ff6f70ab1e93b3c6384e08424a71e97f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-26793) Flink Cassandra connector performance issue

2022-04-15 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-26793?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522355#comment-17522355
 ] 

Etienne Chauchot edited comment on FLINK-26793 at 4/15/22 12:58 PM:


[~bumblebee] I reproduced the behavior you observed: I ran [this streaming 
infinite 
pipeline|https://github.com/echauchot/flink-samples/blob/master/src/main/java/org/example/CassandraPojoSinkStreamingExample.java]
 for 3 days on a local flink 1.14.4 cluster + cassandra 3.0 docker (I did not 
have rights to instanciate a cassandra cluster on Azure). The pipeline has 
checkpointing configured every 10 min with exactly once semantics and no 
watermark defined. It was run at parallelism 16 which corresponds to the number 
of cores on my laptop. I created a source that gives pojos every 100 ms.  The 
source is mono-threaded so at parallelism 1.  See all the screenshots

I ran the pipeline for more than 72 hours and indeed after little less than 
72h, I got an exception from Cassandra cluster see task manager log:

{code:java}
2022-04-13 16:38:15,227 ERROR 
org.apache.flink.streaming.connectors.cassandra.CassandraPojoSink [] - Error 
while sending value.
com.datastax.driver.core.exceptions.WriteTimeoutException: Cassandra timeout 
during write query at consistency LOCAL_ONE (1 replica were required but only 0 
acknowledged the write)
{code}
This exception means that Cassandra coordinator node (internal Cassandra) 
waited too long for an internal replication (raplication to another node in the 
same casssandra "datacenter") and did not ack the write. 

This led to a failure of the write task and to a restoration of the job from 
the last checkpoint see job manager log:

{code:java}
2022-04-13 16:38:20,847 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph   [] - Job Cassandra 
Pojo Sink Streaming example (dc7522bc1855f6f98038ac2b4eed4095) switched from 
state RESTARTING to RUNNING.
2022-04-13 16:38:20,850 INFO  
org.apache.flink.runtime.checkpoint.CheckpointCoordinator[] - Restoring job 
dc7522bc1855f6f98038ac2b4eed4095 from Checkpoint 136 @ 1649858983772 for 
dc7522bc1855f6f98038ac2b4eed4095 located at 
file:/tmp/flink-checkpoints/dc7522bc1855f6f98038ac2b4eed4095/chk-136.
{code}

This restoration led to the restoration of the _CassandraPojoSink_ and to the 
call of _CassandraPojoSink#open_ which reconnects to cassandra cluster and 
re-creates the related _MappingManager_

So in short, this is what I supposed in my previous comments. Restoring from 
checkpoints  slows down your writes (job restart time +  cassandra driver state 
re-creation - reconnection,  prepared statements recreation in the 
MappingManager etc... -)

The problem is that the timeout comes from Cassandra itself not from Flink and 
it is normal that Flink restores the job in such circumstances. 

What you can do is to increase the Cassandra write timeout to adapt to your 
workload in your Cassandra cluster so that such timeout errors do not happen. 
For that you need to raise _write_request_timeout_in_ms_  conf parameter in 
your _cassandra.yml_.

I do not recommend that you lower the replication factor in your Cassandra 
cluster (I did that only for local tests on Flink) because it is mandatory that 
you do not loose data in case of your Cassandra cluster failure. Waiting for a 
single replica for write acknowledge is the minimum level for this guarantee in 
Cassandra.

Best
Etienne

 [^Capture d’écran de 2022-04-14 16-34-59.png]
 [^Capture d’écran de 2022-04-14 16-35-30.png] 
 [^Capture d’écran de 2022-04-14 16-35-07.png] 
 [^jobmanager_log.txt] 
 [^taskmanager_127.0.1.1_33251-af56fa_log] 



was (Author: echauchot):
[~bumblebee] I reproduced the behavior you observed: I ran [this streaming 
infinite 
pipeline|https://github.com/echauchot/flink-samples/blob/master/src/main/java/org/example/CassandraPojoSinkStreamingExample.java]
 for 3 days on a local flink 1.14.4 cluster + cassandra 3.0 docker (I did not 
have rights to instanciate a cassandra cluster on Azure). The pipeline has 
checkpointing configured every 10 min with exactly once semantics and no 
watermark defined. It was run at parallelism 16 which corresponds to the number 
of cores on my laptop. I created a source that gives pojos every 100 ms.  The 
source is mono-threaded so at parallelism 1.  See all the screenshots

I ran the pipeline for more than 72 hours and indeed after little less than 
72h, I got an exception from Cassandra cluster see task manager log:

{code:java}
2022-04-13 16:38:15,227 ERROR 
org.apache.flink.streaming.connectors.cassandra.CassandraPojoSink [] - Error 
while sending value.
com.datastax.driver.core.exceptions.WriteTimeoutException: Cassandra timeout 
during write query at consistency LOCAL_ONE (1 replica were required but only 0 
acknowledged the write)
{code}
This exception means that Cassandra coordinator node (internal Cassandra) 
waited 

[jira] [Comment Edited] (FLINK-26824) Upgrade Flink's supported Cassandra versions to match with the Cassandra community supported versions

2022-04-15 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-26824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17521203#comment-17521203
 ] 

Etienne Chauchot edited comment on FLINK-26824 at 4/15/22 12:57 PM:


and by the way, there was a complete refactoring of datastax cassandra driver 
and a relocation starting at 4.0 that impacts all our code base. But using 
latest 3.x driver works with both Cassandra 4.x and 3.x. So I suggest we keep 
the cassandra-driver 3.x versions


was (Author: echauchot):
and by the way, there was a complete refactoring of cassandra driver and a 
relocation starting at 4.0 that impacts all our code base. But using latest 3.x 
driver works with both Cassandra 4.x and 3.x. So I suggest we keep the 
cassandra-driver 3.x versions

> Upgrade Flink's supported Cassandra versions to match with the Cassandra 
> community supported versions
> -
>
> Key: FLINK-26824
> URL: https://issues.apache.org/jira/browse/FLINK-26824
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Cassandra
>Reporter: Martijn Visser
>Assignee: Etienne Chauchot
>Priority: Major
>
> Flink's Cassandra connector is currently only supporting 
> com.datastax.cassandra:cassandra-driver-core version 3.0.0. 
> The Cassandra community supports 3 versions. One GA (general availability, 
> the latest version), one stable and one older supported release per 
> https://cassandra.apache.org/_/download.html.
> These are currently:
> Cassandra 4.0 (GA)
> Cassandra 3.11 (Stable)
> Cassandra 3.0 (Older supported release).
> We should support (and follow) the supported versions by the Cassandra 
> community



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] huacaicai commented on pull request #19492: Update TimeWindow.java

2022-04-15 Thread GitBox


huacaicai commented on PR #19492:
URL: https://github.com/apache/flink/pull/19492#issuecomment-1100087484

   Based on the rolling window is a special window with a step size equal to 
the window size, so I modified the parameter to the window step size, which can 
better express the meaning of the method parameters. Please pass


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-6755) Allow triggering Checkpoints through command line client

2022-04-15 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-6755?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17458438#comment-17458438
 ] 

Piotr Nowojski edited comment on FLINK-6755 at 4/15/22 12:52 PM:
-

The motivation behind this feature request will be covered by FLINK-25276.

As mentioned above by Aljoscha, there might be still a value of exposing manual 
checkpoint triggering REST API hook, so I'm keeping this ticket open. However 
it doesn't look like such feature is well motivated. Implementation of this 
should be quite straightforward since Flink internally already supports this 
(FLINK-24280). It's just not exposed in anyway to the user.

edit: Although this idea might be still valid, I strongly think we should not 
expose checkpoint directory when triggering checkpoints as proposed in the 
description:
{noformat}
./bin/flink checkpoint  [checkpointDirectory]
{noformat}
Since checkpoints are owned fully by Flink, CLI/REST API call to trigger 
checkpoints should not expose anything like that. If anything, it should be 
just a simple trigger with optionally parameters like whether the checkpoint 
should be full or incremental.

The same remark applies to:
{noformat}
./bin/flink cancel -c [targetDirectory] 
{nofrmat}
and I don't see a point of supporting cancelling/stopping job with checkpoint.



was (Author: pnowojski):
The motivation behind this feature request will be covered by FLINK-25276.

As mentioned above by Aljoscha, there might be still a value of exposing manual 
checkpoint triggering REST API hook, so I'm keeping this ticket open. However 
it doesn't look like such feature is well motivated. Implementation of this 
should be quite straightforward since Flink internally already supports this 
(FLINK-24280). It's just not exposed in anyway to the user.

edit: Although this idea might be still valid, I strongly think we should not 
expose checkpoint directory when triggering checkpoints as proposed in the 
description:
{noformat}
./bin/flink checkpoint  [checkpointDirectory]
{noformat}
Since checkpoints are owned fully by Flink, CLI/REST API call to trigger 
checkpoints should not expose anything like that. If anything, it should be 
just a simple trigger with optionally parameters like whether the checkpoint 
should be full or incremental.


> Allow triggering Checkpoints through command line client
> 
>
> Key: FLINK-6755
> URL: https://issues.apache.org/jira/browse/FLINK-6755
> Project: Flink
>  Issue Type: New Feature
>  Components: Command Line Client, Runtime / Checkpointing
>Affects Versions: 1.3.0
>Reporter: Gyula Fora
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-unassigned
>
> The command line client currently only allows triggering (and canceling with) 
> Savepoints. 
> While this is good if we want to fork or modify the pipelines in a 
> non-checkpoint compatible way, now with incremental checkpoints this becomes 
> wasteful for simple job restarts/pipeline updates. 
> I suggest we add a new command: 
> ./bin/flink checkpoint  [checkpointDirectory]
> and a new flag -c for the cancel command to indicate we want to trigger a 
> checkpoint:
> ./bin/flink cancel -c [targetDirectory] 
> Otherwise this can work similar to the current savepoint taking logic, we 
> could probably even piggyback on the current messages by adding boolean flag 
> indicating whether it should be a savepoint or a checkpoint.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Created] (FLINK-27266) in hive_read_write.md, there is an extra backtick in the description

2022-04-15 Thread Jira
陈磊 created FLINK-27266:
--

 Summary: in hive_read_write.md, there is an extra backtick in the 
description 
 Key: FLINK-27266
 URL: https://issues.apache.org/jira/browse/FLINK-27266
 Project: Flink
  Issue Type: New Feature
  Components: Documentation
Reporter: 陈磊
 Attachments: image-2022-04-15-20-52-40-819.png

in hive_read_write.md, there is an extra backtick in the description 
!image-2022-04-15-20-52-40-819.png!



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] huacaicai commented on pull request #19492: Update TimeWindow.java

2022-04-15 Thread GitBox


huacaicai commented on PR #19492:
URL: https://github.com/apache/flink/pull/19492#issuecomment-1100087060

   The window size cannot accurately express the true meaning of the parameter, 
because the sliding window needs to pass in the step size, and the rolling 
window passes in the window size. Based on the rolling window is a special 
window with a step size equal to the window size, so I modified the parameter 
to the window step size, which can better express the meaning of the method 
parameters


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] huacaicai commented on pull request #19492: Update TimeWindow.java

2022-04-15 Thread GitBox


huacaicai commented on PR #19492:
URL: https://github.com/apache/flink/pull/19492#issuecomment-1100086913

   The window size cannot accurately express the true meaning of the parameter, 
because the sliding window needs to pass in the step size, and the rolling 
window passes in the window size. Based on the rolling window is a special 
window with a step size equal to the window size, so I modified the parameter 
to the window step size, which can better express the meaning of the method 
parameters.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-6755) Allow triggering Checkpoints through command line client

2022-04-15 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-6755?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17458438#comment-17458438
 ] 

Piotr Nowojski edited comment on FLINK-6755 at 4/15/22 12:51 PM:
-

The motivation behind this feature request will be covered by FLINK-25276.

As mentioned above by Aljoscha, there might be still a value of exposing manual 
checkpoint triggering REST API hook, so I'm keeping this ticket open. However 
it doesn't look like such feature is well motivated. Implementation of this 
should be quite straightforward since Flink internally already supports this 
(FLINK-24280). It's just not exposed in anyway to the user.

edit: Although this idea might be still valid, I strongly think we should not 
expose checkpoint directory when triggering checkpoints as proposed in the 
description:
{noformat}
./bin/flink checkpoint  [checkpointDirectory]
{noformat}
Since checkpoints are owned fully by Flink, CLI/REST API call to trigger 
checkpoints should not expose anything like that. If anything, it should be 
just a simple trigger with optionally parameters like whether the checkpoint 
should be full or incremental.



was (Author: pnowojski):
The motivation behind this feature request will be covered by FLINK-25276.

As mentioned above by Aljoscha, there might be still a value of exposing manual 
checkpoint triggering REST API hook, so I'm keeping this ticket open. However 
it doesn't look like such feature is well motivated. Implementation of this 
should be quite straightforward since Flink internally already supports this 
(FLINK-24280). It's just not exposed in anyway to the user.

> Allow triggering Checkpoints through command line client
> 
>
> Key: FLINK-6755
> URL: https://issues.apache.org/jira/browse/FLINK-6755
> Project: Flink
>  Issue Type: New Feature
>  Components: Command Line Client, Runtime / Checkpointing
>Affects Versions: 1.3.0
>Reporter: Gyula Fora
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-unassigned
>
> The command line client currently only allows triggering (and canceling with) 
> Savepoints. 
> While this is good if we want to fork or modify the pipelines in a 
> non-checkpoint compatible way, now with incremental checkpoints this becomes 
> wasteful for simple job restarts/pipeline updates. 
> I suggest we add a new command: 
> ./bin/flink checkpoint  [checkpointDirectory]
> and a new flag -c for the cancel command to indicate we want to trigger a 
> checkpoint:
> ./bin/flink cancel -c [targetDirectory] 
> Otherwise this can work similar to the current savepoint taking logic, we 
> could probably even piggyback on the current messages by adding boolean flag 
> indicating whether it should be a savepoint or a checkpoint.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] huacaicai opened a new pull request, #19492: Update TimeWindow.java

2022-04-15 Thread GitBox


huacaicai opened a new pull request, #19492:
URL: https://github.com/apache/flink/pull/19492

   The window size cannot accurately express the true meaning of the parameter, 
because the sliding window needs to pass in the step size, and the rolling 
window passes in the window size. Based on the rolling window is a special 
window with a step size equal to the window size, so I modified the parameter 
to the window step size, which can better express the meaning of the method 
parameters.
   
   
   
   ## What is the purpose of the change
   
   *(For example: This pull request makes task deployment go through the blob 
server, rather than through RPC. That way we avoid re-transferring them on each 
deployment (during recovery).)*
   
   
   ## Brief change log
   
   *(for example:)*
 - *The TaskInfo is stored in the blob store on job creation time as a 
persistent artifact*
 - *Deployments RPC transmits only the blob storage reference*
 - *TaskManagers retrieve the TaskInfo from the blob cache*
   
   
   ## Verifying this change
   
   Please make sure both new and modified tests in this PR follows the 
conventions defined in our code quality guide: 
https://flink.apache.org/contributing/code-style-and-quality-common.html#testing
   
   *(Please pick either of the following options)*
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   *(or)*
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   *(or)*
   
   This change added tests and can be verified as follows:
   
   *(example:)*
 - *Added integration tests for end-to-end deployment with large payloads 
(100MB)*
 - *Extended integration test for recovery after master (JobManager) 
failure*
 - *Added test that validates that TaskInfo is transferred only once across 
recoveries*
 - *Manually verified the change by running a 4 node cluster with 2 
JobManagers and 4 TaskManagers, a stateful streaming program, and killing one 
JobManager and two TaskManagers during the execution, verifying that recovery 
happens correctly.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / no)
 - The serializers: (yes / no / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / no / 
don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know)
 - The S3 file system connector: (yes / no / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-27101) Periodically break the chain of incremental checkpoint

2022-04-15 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522815#comment-17522815
 ] 

Piotr Nowojski edited comment on FLINK-27101 at 4/15/22 12:48 PM:
--

Checkpoint would be still owned by the Flink itself. Only triggering mechanism 
would be exposed. In principle it's not far of from externally induced 
checkpoints like [1] or [2]. Does it matter if checkpoints are triggered via 
some internal timer or a REST API call? So all in all I don't see any problems 
here, but please correct me if I'm wrong.

[1] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/streaming/api/checkpoint/ExternallyInducedSource.CheckpointTrigger.html
[2] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/api/connector/source/ExternallyInducedSourceReader.html

edit: just note that the API as initially proposed in FLINK-6755 is clearly 
incompatible with the checkpoint ownership model. FLINK-6755 is pretty old, and 
in the description it was proposed to pass "checkpoint directory" parameter, 
which we definitely should not do. 


was (Author: pnowojski):
Checkpoint would be still owned by the Flink itself. Only triggering mechanism 
would be exposed. In principle it's not far of from externally induced 
checkpoints like [1] or [2]. Does it matter if checkpoints are triggered via 
some internal timer or a REST API call? So all in all I don't see any problems 
here, but please correct me if I'm wrong.

[1] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/streaming/api/checkpoint/ExternallyInducedSource.CheckpointTrigger.html
[2] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/api/connector/source/ExternallyInducedSourceReader.html

> Periodically break the chain of incremental checkpoint
> --
>
> Key: FLINK-27101
> URL: https://issues.apache.org/jira/browse/FLINK-27101
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / Checkpointing
>Reporter: Steven Zhen Wu
>Priority: Major
>
> Incremental checkpoint is almost a must for large-state jobs. It greatly 
> reduces the bytes uploaded to DFS per checkpoint. However, there are  a few 
> implications from incremental checkpoint that are problematic for production 
> operations.  Will use S3 as an example DFS for the rest of description.
> 1. Because there is no way to deterministically know how far back the 
> incremental checkpoint can refer to files uploaded to S3, it is very 
> difficult to set S3 bucket/object TTL. In one application, we have observed 
> Flink checkpoint referring to files uploaded over 6 months ago. S3 TTL can 
> corrupt the Flink checkpoints.
> S3 TTL is important for a few reasons
> - purge orphaned files (like external checkpoints from previous deployments) 
> to keep the storage cost in check. This problem can be addressed by 
> implementing proper garbage collection (similar to JVM) by traversing the 
> retained checkpoints from all jobs and traverse the file references. But that 
> is an expensive solution from engineering cost perspective.
> - Security and privacy. E.g., there may be requirement that Flink state can't 
> keep the data for more than some duration threshold (hours/days/weeks). 
> Application is expected to purge keys to satisfy the requirement. However, 
> with incremental checkpoint and how deletion works in RocksDB, it is hard to 
> set S3 TTL to purge S3 files. Even though those old S3 files don't contain 
> live keys, they may still be referrenced by retained Flink checkpoints.
> 2. Occasionally, corrupted checkpoint files (on S3) are observed. As a 
> result, restoring from checkpoint failed. With incremental checkpoint, it 
> usually doesn't help to try other older checkpoints, because they may refer 
> to the same corrupted file. It is unclear whether the corruption happened 
> before or during S3 upload. This risk can be mitigated with periodical 
> savepoints.
> It all boils down to periodical full snapshot (checkpoint or savepoint) to 
> deterministically break the chain of incremental checkpoints. Search the jira 
> history, the behavior that FLINK-23949 [1] trying to fix is actually close to 
> what we would need here.
> There are a few options
> 1. Periodically trigger savepoints (via control plane). This is actually not 
> a bad practice and might be appealing to some people. The problem is that it 
> requires a job deployment to break the chain of incremental checkpoint. 
> periodical job deployment may sound hacky. If we make the behavior of full 
> checkpoint after a savepoint (fixed in FLINK-23949) configurable, it might be 
> an acceptable compromise. The benefit is that no job deployment is 

[jira] [Comment Edited] (FLINK-6755) Allow triggering Checkpoints through command line client

2022-04-15 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-6755?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17458438#comment-17458438
 ] 

Piotr Nowojski edited comment on FLINK-6755 at 4/15/22 12:48 PM:
-

The motivation behind this feature request will be covered by FLINK-25276.

As mentioned above by Aljoscha, there might be still a value of exposing manual 
checkpoint triggering REST API hook, so I'm keeping this ticket open. However 
it doesn't look like such feature is well motivated. Implementation of this 
should be quite straightforward since Flink internally already supports this 
(FLINK-24280). It's just not exposed in anyway to the user.


was (Author: pnowojski):
The motivation behind this feature request will be covered by FLINK-25276.

As mentioned above by Aljoscha, there might be still a value of exposing manual 
checkpoint triggering REST API hook, so I'm keeping this ticket open. However 
it doesn't look like such feature is well motivated. Implementation of this 
should be quite straightforward since Flink internally already supports this 
(FLINK-24280). It's just not exposed in anyway to the user.

> Allow triggering Checkpoints through command line client
> 
>
> Key: FLINK-6755
> URL: https://issues.apache.org/jira/browse/FLINK-6755
> Project: Flink
>  Issue Type: New Feature
>  Components: Command Line Client, Runtime / Checkpointing
>Affects Versions: 1.3.0
>Reporter: Gyula Fora
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-unassigned
>
> The command line client currently only allows triggering (and canceling with) 
> Savepoints. 
> While this is good if we want to fork or modify the pipelines in a 
> non-checkpoint compatible way, now with incremental checkpoints this becomes 
> wasteful for simple job restarts/pipeline updates. 
> I suggest we add a new command: 
> ./bin/flink checkpoint  [checkpointDirectory]
> and a new flag -c for the cancel command to indicate we want to trigger a 
> checkpoint:
> ./bin/flink cancel -c [targetDirectory] 
> Otherwise this can work similar to the current savepoint taking logic, we 
> could probably even piggyback on the current messages by adding boolean flag 
> indicating whether it should be a savepoint or a checkpoint.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Comment Edited] (FLINK-27101) Periodically break the chain of incremental checkpoint

2022-04-15 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522815#comment-17522815
 ] 

Piotr Nowojski edited comment on FLINK-27101 at 4/15/22 12:47 PM:
--

Checkpoint would be still owned by the Flink itself. Only triggering mechanism 
would be exposed. In principle it's not far of from externally induced 
checkpoints like [1] or [2]. Does it matter if checkpoints are triggered via 
some internal timer or a REST API call? So all in all I don't see any problems 
here, but please correct me if I'm wrong.

[1] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/streaming/api/checkpoint/ExternallyInducedSource.CheckpointTrigger.html
[2] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/api/connector/source/ExternallyInducedSourceReader.html


was (Author: pnowojski):
Checkpoint would be still owned by the Flink itself. Only triggering mechanism 
would be exposed. In principle it's not far of from externally induced 
checkpoints like [1] or [2]. Does it matter if checkpoints are triggered via 
some internal timer or a REST API call? So all in all I don't see any problems 
here, but please correct me if I'm wrong.

[1] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/streaming/api/checkpoint/ExternallyInducedSource.CheckpointTrigger.html
[2] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/api/connector/source/ExternallyInducedSourceReader.html

> Periodically break the chain of incremental checkpoint
> --
>
> Key: FLINK-27101
> URL: https://issues.apache.org/jira/browse/FLINK-27101
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / Checkpointing
>Reporter: Steven Zhen Wu
>Priority: Major
>
> Incremental checkpoint is almost a must for large-state jobs. It greatly 
> reduces the bytes uploaded to DFS per checkpoint. However, there are  a few 
> implications from incremental checkpoint that are problematic for production 
> operations.  Will use S3 as an example DFS for the rest of description.
> 1. Because there is no way to deterministically know how far back the 
> incremental checkpoint can refer to files uploaded to S3, it is very 
> difficult to set S3 bucket/object TTL. In one application, we have observed 
> Flink checkpoint referring to files uploaded over 6 months ago. S3 TTL can 
> corrupt the Flink checkpoints.
> S3 TTL is important for a few reasons
> - purge orphaned files (like external checkpoints from previous deployments) 
> to keep the storage cost in check. This problem can be addressed by 
> implementing proper garbage collection (similar to JVM) by traversing the 
> retained checkpoints from all jobs and traverse the file references. But that 
> is an expensive solution from engineering cost perspective.
> - Security and privacy. E.g., there may be requirement that Flink state can't 
> keep the data for more than some duration threshold (hours/days/weeks). 
> Application is expected to purge keys to satisfy the requirement. However, 
> with incremental checkpoint and how deletion works in RocksDB, it is hard to 
> set S3 TTL to purge S3 files. Even though those old S3 files don't contain 
> live keys, they may still be referrenced by retained Flink checkpoints.
> 2. Occasionally, corrupted checkpoint files (on S3) are observed. As a 
> result, restoring from checkpoint failed. With incremental checkpoint, it 
> usually doesn't help to try other older checkpoints, because they may refer 
> to the same corrupted file. It is unclear whether the corruption happened 
> before or during S3 upload. This risk can be mitigated with periodical 
> savepoints.
> It all boils down to periodical full snapshot (checkpoint or savepoint) to 
> deterministically break the chain of incremental checkpoints. Search the jira 
> history, the behavior that FLINK-23949 [1] trying to fix is actually close to 
> what we would need here.
> There are a few options
> 1. Periodically trigger savepoints (via control plane). This is actually not 
> a bad practice and might be appealing to some people. The problem is that it 
> requires a job deployment to break the chain of incremental checkpoint. 
> periodical job deployment may sound hacky. If we make the behavior of full 
> checkpoint after a savepoint (fixed in FLINK-23949) configurable, it might be 
> an acceptable compromise. The benefit is that no job deployment is required 
> after savepoints.
> 2. Build the feature in Flink incremental checkpoint. Periodically (with some 
> cron style config) trigger a full checkpoint to break the incremental chain. 
> If the full checkpoint failed (due to whatever reason), the following 
> checkpoints 

[GitHub] [flink] flinkbot commented on pull request #19491: [hotfix][docs-zh] Delete the "mongodb.md" file under "Formats" under "DataStream Connectors".

2022-04-15 Thread GitBox


flinkbot commented on PR #19491:
URL: https://github.com/apache/flink/pull/19491#issuecomment-1100083966

   
   ## CI report:
   
   * 4523dd09c135ef477ff1698f8a4bae96ab3e4c20 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] liuzhuang2017 commented on pull request #19491: [hotfix][docs-zh] Delete the "mongodb.md" file under "Formats" under "DataStream Connectors".

2022-04-15 Thread GitBox


liuzhuang2017 commented on PR #19491:
URL: https://github.com/apache/flink/pull/19491#issuecomment-1100083881

   @wuchong ,Hi,please help me review  this pr when you are free time, thank 
you.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-27101) Periodically break the chain of incremental checkpoint

2022-04-15 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522815#comment-17522815
 ] 

Piotr Nowojski commented on FLINK-27101:


Checkpoint would be still owned by the Flink itself. Only triggering mechanism 
would be exposed. In principle it's not far of from externally induced 
checkpoints like [1] or [2]. Does it matter if checkpoints are triggered via 
some internal timer or a REST API call? So all in all I don't see any problems 
here, but please correct me if I'm wrong.

[1] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/streaming/api/checkpoint/ExternallyInducedSource.CheckpointTrigger.html
[2] 
https://nightlies.apache.org/flink/flink-docs-release-1.15/api/java/org/apache/flink/api/connector/source/ExternallyInducedSourceReader.html

> Periodically break the chain of incremental checkpoint
> --
>
> Key: FLINK-27101
> URL: https://issues.apache.org/jira/browse/FLINK-27101
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / Checkpointing
>Reporter: Steven Zhen Wu
>Priority: Major
>
> Incremental checkpoint is almost a must for large-state jobs. It greatly 
> reduces the bytes uploaded to DFS per checkpoint. However, there are  a few 
> implications from incremental checkpoint that are problematic for production 
> operations.  Will use S3 as an example DFS for the rest of description.
> 1. Because there is no way to deterministically know how far back the 
> incremental checkpoint can refer to files uploaded to S3, it is very 
> difficult to set S3 bucket/object TTL. In one application, we have observed 
> Flink checkpoint referring to files uploaded over 6 months ago. S3 TTL can 
> corrupt the Flink checkpoints.
> S3 TTL is important for a few reasons
> - purge orphaned files (like external checkpoints from previous deployments) 
> to keep the storage cost in check. This problem can be addressed by 
> implementing proper garbage collection (similar to JVM) by traversing the 
> retained checkpoints from all jobs and traverse the file references. But that 
> is an expensive solution from engineering cost perspective.
> - Security and privacy. E.g., there may be requirement that Flink state can't 
> keep the data for more than some duration threshold (hours/days/weeks). 
> Application is expected to purge keys to satisfy the requirement. However, 
> with incremental checkpoint and how deletion works in RocksDB, it is hard to 
> set S3 TTL to purge S3 files. Even though those old S3 files don't contain 
> live keys, they may still be referrenced by retained Flink checkpoints.
> 2. Occasionally, corrupted checkpoint files (on S3) are observed. As a 
> result, restoring from checkpoint failed. With incremental checkpoint, it 
> usually doesn't help to try other older checkpoints, because they may refer 
> to the same corrupted file. It is unclear whether the corruption happened 
> before or during S3 upload. This risk can be mitigated with periodical 
> savepoints.
> It all boils down to periodical full snapshot (checkpoint or savepoint) to 
> deterministically break the chain of incremental checkpoints. Search the jira 
> history, the behavior that FLINK-23949 [1] trying to fix is actually close to 
> what we would need here.
> There are a few options
> 1. Periodically trigger savepoints (via control plane). This is actually not 
> a bad practice and might be appealing to some people. The problem is that it 
> requires a job deployment to break the chain of incremental checkpoint. 
> periodical job deployment may sound hacky. If we make the behavior of full 
> checkpoint after a savepoint (fixed in FLINK-23949) configurable, it might be 
> an acceptable compromise. The benefit is that no job deployment is required 
> after savepoints.
> 2. Build the feature in Flink incremental checkpoint. Periodically (with some 
> cron style config) trigger a full checkpoint to break the incremental chain. 
> If the full checkpoint failed (due to whatever reason), the following 
> checkpoints should attempt full checkpoint as well until one successful full 
> checkpoint is completed.
> 3. For the security/privacy requirement, the main thing is to apply 
> compaction on the deleted keys. That could probably avoid references to the 
> old files. Is there any RocksDB compation can achieve full compaction of 
> removing old delete markers. Recent delete markers are fine
> [1] https://issues.apache.org/jira/browse/FLINK-23949



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Commented] (FLINK-27265) The problem described by the filesystem.md documentation

2022-04-15 Thread Jira


[ 
https://issues.apache.org/jira/browse/FLINK-27265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522814#comment-17522814
 ] 

陈磊 commented on FLINK-27265:


[https://github.com/apache/flink/pull/19489]

[~MartijnVisser] Hi, Could you merge this PR? Thanks.

> The problem described by the filesystem.md documentation
> 
>
> Key: FLINK-27265
> URL: https://issues.apache.org/jira/browse/FLINK-27265
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: 陈磊
>Priority: Minor
> Attachments: image-2022-04-15-20-25-26-150.png
>
>
> in filesystem.md, flink comes with four built-in BulkWriter factories, in 
> fact, the list has five.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] liuzhuang2017 opened a new pull request, #19491: [hotfix][docs-zh] Delete the "mongodb.md" file under "Formats" under "DataStream Connectors".

2022-04-15 Thread GitBox


liuzhuang2017 opened a new pull request, #19491:
URL: https://github.com/apache/flink/pull/19491

   
   ## What is the purpose of the change
   
   #  This is the English document:
   
   
![image](https://user-images.githubusercontent.com/95120044/163571308-5a5a6d50-e993-4088-9934-478838c2b98f.png)
   
   # This is the Chinese document: 
   
   
![image](https://user-images.githubusercontent.com/95120044/163571468-910101ff-fe9d-4158-a01a-b047ed0a7284.png)
   
   **As we can see from the above figure, the "mongdb.md" file in the Chinese 
document is redundant compared to the English document.**
   
   
   ## Brief change log
   
   Delete the "mongodb.md" file under "Formats" under "DataStream Connectors".
   
   ## Verifying this change
   
   Delete the "mongodb.md" file under "Formats" under "DataStream Connectors".


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] luoyuxia commented on pull request #19435: [FLINK-27121][docs-zh]Translate "Configuration#overview" paragraph and the code example in "Application Development > Table API & SQL" to Ch

2022-04-15 Thread GitBox


luoyuxia commented on PR #19435:
URL: https://github.com/apache/flink/pull/19435#issuecomment-1100081624

   @leizhou666 Thanks for you contribution. LGTM
   @MartijnVisser Could you please help merge?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-27265) The problem described by the filesystem.md documentation

2022-04-15 Thread Jira


 [ 
https://issues.apache.org/jira/browse/FLINK-27265?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

陈磊 updated FLINK-27265:
---
Description: in filesystem.md, flink comes with four built-in BulkWriter 
factories, in fact, the list has five.  (was: 1)  in hive_read_write.md, there 
is an extra backtick in the description

2) in filesystem.md, the format parameter described is incomplete.

3) in filesystem.md, flink comes with four built-in BulkWriter factories, in 
fact, the list has five.)
Summary: The problem described by the filesystem.md documentation  
(was: The problem described by the filesystem.md and 
hive_write_read.mddocumentation)

> The problem described by the filesystem.md documentation
> 
>
> Key: FLINK-27265
> URL: https://issues.apache.org/jira/browse/FLINK-27265
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: 陈磊
>Priority: Minor
> Attachments: image-2022-04-15-20-25-26-150.png
>
>
> in filesystem.md, flink comes with four built-in BulkWriter factories, in 
> fact, the list has five.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-27265) The problem described by the filesystem.md documentation

2022-04-15 Thread Jira


 [ 
https://issues.apache.org/jira/browse/FLINK-27265?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

陈磊 updated FLINK-27265:
---
Attachment: (was: image-2022-04-15-20-26-09-895.png)

> The problem described by the filesystem.md documentation
> 
>
> Key: FLINK-27265
> URL: https://issues.apache.org/jira/browse/FLINK-27265
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: 陈磊
>Priority: Minor
> Attachments: image-2022-04-15-20-25-26-150.png
>
>
> in filesystem.md, flink comes with four built-in BulkWriter factories, in 
> fact, the list has five.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-27265) The problem described by the filesystem.md documentation

2022-04-15 Thread Jira


 [ 
https://issues.apache.org/jira/browse/FLINK-27265?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

陈磊 updated FLINK-27265:
---
Attachment: (was: image-2022-04-15-20-20-00-116.png)

> The problem described by the filesystem.md documentation
> 
>
> Key: FLINK-27265
> URL: https://issues.apache.org/jira/browse/FLINK-27265
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Reporter: 陈磊
>Priority: Minor
> Attachments: image-2022-04-15-20-25-26-150.png
>
>
> in filesystem.md, flink comes with four built-in BulkWriter factories, in 
> fact, the list has five.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Commented] (FLINK-27155) Reduce multiple reads to the same Changelog file in the same taskmanager during restore

2022-04-15 Thread Feifan Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522810#comment-17522810
 ] 

Feifan Wang commented on FLINK-27155:
-

Hi [~roman] , I think we can achieve this by :
 * Chang{color:#172b4d}e _StateChangeFormat#read(StreamStateHandle, long)_ to 
_StateChangeFormat#read(DataInputStream)_{color}

 * Add new class _StateChangeIteratorWithLocalCache_ implements 
{_}StateChangelogHandleStreamHandleReader.StateChangeIterator{_}. It downloads 
the file referenced by _StreamStateHandle_ from DFS, uncompress it and save in 
a local temporary file. Then, open a _FileInputStream_ on the local temporary 
file and seek to the offset before pass it to 
{_}StateChangeFormat#read(DataInputStream){_}.  The local file is kept for a 
few minutes (maybe should expose an option to user), during which time all 
requests to the same _StreamStateHandle_ are carried by the local file.

How do you think about ?

> Reduce multiple reads to the same Changelog file in the same taskmanager 
> during restore
> ---
>
> Key: FLINK-27155
> URL: https://issues.apache.org/jira/browse/FLINK-27155
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / State Backends
>Reporter: Feifan Wang
>Assignee: Feifan Wang
>Priority: Major
> Fix For: 1.16.0
>
>
> h3. Background
> In the current implementation, State changes of different operators in the 
> same taskmanager may be written to the same changelog file, which effectively 
> reduces the number of files and requests to DFS.
> But on the other hand, the current implementation also reads the same 
> changelog file multiple times on recovery. More specifically, the number of 
> times the same changelog file is accessed is related to the number of 
> ChangeSets contained in it. And since each read needs to skip the preceding 
> bytes, this network traffic is also wasted.
> The result is a lot of unnecessary request to DFS when there are multiple 
> slots and keyed state in the same taskmanager.
> h3. Proposal
> We can reduce multiple reads to the same changelog file in the same 
> taskmanager during restore.
> One possible approach is to read the changelog file all at once and cache it 
> in memory or local file for a period of time when reading the changelog file.
> I think this could be a subtask of [v2 FLIP-158: Generalized incremental 
> checkpoints|https://issues.apache.org/jira/browse/FLINK-25842] .
> Hi [~ym] , [~roman]  how do you think about ?



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Created] (FLINK-27265) The problem described by the filesystem.md and hive_write_read.mddocumentation

2022-04-15 Thread Jira
陈磊 created FLINK-27265:
--

 Summary: The problem described by the filesystem.md and 
hive_write_read.mddocumentation
 Key: FLINK-27265
 URL: https://issues.apache.org/jira/browse/FLINK-27265
 Project: Flink
  Issue Type: Bug
  Components: Documentation
Reporter: 陈磊
 Attachments: image-2022-04-15-20-20-00-116.png, 
image-2022-04-15-20-25-26-150.png, image-2022-04-15-20-26-09-895.png

1)  in hive_read_write.md, there is an extra backtick in the description

2) in filesystem.md, the format parameter described is incomplete.

3) in filesystem.md, flink comes with four built-in BulkWriter factories, in 
fact, the list has five.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Assigned] (FLINK-27264) Add ITCase for concurrent batch overwrite and streaming insert

2022-04-15 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee reassigned FLINK-27264:


Assignee: Jane Chan

> Add ITCase for concurrent batch overwrite and streaming insert
> --
>
> Key: FLINK-27264
> URL: https://issues.apache.org/jira/browse/FLINK-27264
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table Store
>Affects Versions: table-store-0.1.0
>Reporter: Jane Chan
>Assignee: Jane Chan
>Priority: Major
> Fix For: table-store-0.1.0
>
> Attachments: image-2022-04-15-19-26-09-649.png
>
>
> !image-2022-04-15-19-26-09-649.png|width=609,height=241!
> Add it case for user story



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Commented] (FLINK-27212) Failed to CAST('abcde', VARBINARY)

2022-04-15 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522807#comment-17522807
 ] 

Jark Wu commented on FLINK-27212:
-

Hi [~matriv], {{x'45F0AB'}} is a {{BINARY}} not {{CHAR/VARCHAR}} type. So, 
{{CAST(x'45F0AB' AS VARBINARY(6))}} is a CAST BINARY to BINRY,  conversion (the 
hex string of course should be in hex digits). Therefore, what we are missing 
is CAST CHAR to BINARY, and this is supported in previous versions and other 
database systems. 

I would still vote this to be a blocker issue. 

cc [~twalthr]

> Failed to CAST('abcde', VARBINARY)
> --
>
> Key: FLINK-27212
> URL: https://issues.apache.org/jira/browse/FLINK-27212
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.16.0
>Reporter: Shengkai Fang
>Assignee: Marios Trivyzas
>Priority: Blocker
> Fix For: 1.16.0
>
>
> Please add test in the CalcITCase
> {code:scala}
> @Test
>   def testCalc(): Unit = {
> val sql =
>   """
> |SELECT CAST('abcde' AS VARBINARY(6))
> |""".stripMargin
> val result = tEnv.executeSql(sql)
> print(result.getResolvedSchema)
> result.print()
>   }
> {code}
> The exception is 
> {code:java}
> Caused by: org.apache.flink.table.api.TableException: Odd number of 
> characters.
>   at 
> org.apache.flink.table.utils.EncodingUtils.decodeHex(EncodingUtils.java:203)
>   at StreamExecCalc$33.processElement(Unknown Source)
>   at 
> org.apache.flink.streaming.runtime.tasks.ChainingOutput.pushToOperator(ChainingOutput.java:99)
>   at 
> org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:80)
>   at 
> org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:39)
>   at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:56)
>   at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:29)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollect(StreamSourceContexts.java:418)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collect(StreamSourceContexts.java:513)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$SwitchingOnClose.collect(StreamSourceContexts.java:103)
>   at 
> org.apache.flink.streaming.api.functions.source.InputFormatSourceFunction.run(InputFormatSourceFunction.java:92)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:67)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:332)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Comment Edited] (FLINK-27212) Failed to CAST('abcde', VARBINARY)

2022-04-15 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522807#comment-17522807
 ] 

Jark Wu edited comment on FLINK-27212 at 4/15/22 12:13 PM:
---

Hi [~matriv], {{x'45F0AB'}} is a {{BINARY}} not {{CHAR/VARCHAR}} type [1]. So, 
{{CAST(x'45F0AB' AS VARBINARY(6))}} is a CAST BINARY to BINRY,  conversion (the 
hex string of course should be in hex digits). Therefore, what we are missing 
is CAST CHAR to BINARY, and this is supported in previous versions and other 
database systems. 

I would still vote this to be a blocker issue. 

cc [~twalthr]

[1]: https://calcite.apache.org/docs/reference.html#scalar-types


was (Author: jark):
Hi [~matriv], {{x'45F0AB'}} is a {{BINARY}} not {{CHAR/VARCHAR}} type. So, 
{{CAST(x'45F0AB' AS VARBINARY(6))}} is a CAST BINARY to BINRY,  conversion (the 
hex string of course should be in hex digits). Therefore, what we are missing 
is CAST CHAR to BINARY, and this is supported in previous versions and other 
database systems. 

I would still vote this to be a blocker issue. 

cc [~twalthr]

> Failed to CAST('abcde', VARBINARY)
> --
>
> Key: FLINK-27212
> URL: https://issues.apache.org/jira/browse/FLINK-27212
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.16.0
>Reporter: Shengkai Fang
>Assignee: Marios Trivyzas
>Priority: Blocker
> Fix For: 1.16.0
>
>
> Please add test in the CalcITCase
> {code:scala}
> @Test
>   def testCalc(): Unit = {
> val sql =
>   """
> |SELECT CAST('abcde' AS VARBINARY(6))
> |""".stripMargin
> val result = tEnv.executeSql(sql)
> print(result.getResolvedSchema)
> result.print()
>   }
> {code}
> The exception is 
> {code:java}
> Caused by: org.apache.flink.table.api.TableException: Odd number of 
> characters.
>   at 
> org.apache.flink.table.utils.EncodingUtils.decodeHex(EncodingUtils.java:203)
>   at StreamExecCalc$33.processElement(Unknown Source)
>   at 
> org.apache.flink.streaming.runtime.tasks.ChainingOutput.pushToOperator(ChainingOutput.java:99)
>   at 
> org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:80)
>   at 
> org.apache.flink.streaming.runtime.tasks.ChainingOutput.collect(ChainingOutput.java:39)
>   at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:56)
>   at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:29)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollect(StreamSourceContexts.java:418)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collect(StreamSourceContexts.java:513)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$SwitchingOnClose.collect(StreamSourceContexts.java:103)
>   at 
> org.apache.flink.streaming.api.functions.source.InputFormatSourceFunction.run(InputFormatSourceFunction.java:92)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:67)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:332)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] luoyuxia commented on a diff in pull request #19419: [FLINK-22317][table] Support DROP column/constraint/watermark for ALTER TABLE statement

2022-04-15 Thread GitBox


luoyuxia commented on code in PR #19419:
URL: https://github.com/apache/flink/pull/19419#discussion_r848239830


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/utils/OperationConverterUtils.java:
##
@@ -117,6 +128,193 @@ public static Operation convertAddReplaceColumns(
 catalogTable.getComment()));
 }
 
+public static Operation convertAlterTableDropConstraint(
+ObjectIdentifier tableIdentifier,
+CatalogTable catalogTable,
+SqlAlterTableDropConstraint alterTableDropConstraint) {
+boolean isPrimaryKey = alterTableDropConstraint.isPrimaryKey();
+Optional oriPrimaryKey =
+catalogTable.getUnresolvedSchema().getPrimaryKey();
+// validate primary key is exists in table
+if (!oriPrimaryKey.isPresent()) {
+throw new ValidationException(
+String.format("Table %s does not exist primary key.", 
tableIdentifier));
+}
+
+String constraintName = null;
+if (alterTableDropConstraint.getConstraintName().isPresent()) {
+constraintName = 
alterTableDropConstraint.getConstraintName().get().getSimple();
+}
+if (!StringUtils.isNullOrWhitespaceOnly(constraintName)
+&& 
!oriPrimaryKey.get().getConstraintName().equals(constraintName)) {

Review Comment:
   Some confuses about the code. Seems the constraints will always be 
primarykey?



##
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/operations/ddl/AlterTableDropConstraintOperation.java:
##
@@ -20,24 +20,42 @@
 
 import org.apache.flink.table.catalog.ObjectIdentifier;
 
+import javax.annotation.Nullable;
+
+import java.util.Optional;
+
 /** Operation of "ALTER TABLE ADD [CONSTRAINT constraintName] ..." clause. * */
 public class AlterTableDropConstraintOperation extends AlterTableOperation {
-private final String constraintName;
+
+private final boolean isPrimaryKey;
+private final @Nullable String constraintName;
 
 public AlterTableDropConstraintOperation(
-ObjectIdentifier tableIdentifier, String constraintName) {
+ObjectIdentifier tableIdentifier,
+boolean isPrimaryKey,

Review Comment:
   Will it better to wrap `isPrimaryKey` and `constraintName` to a class may 
called `Constraint`? 



##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/utils/OperationConverterUtils.java:
##
@@ -117,6 +128,193 @@ public static Operation convertAddReplaceColumns(
 catalogTable.getComment()));
 }
 
+public static Operation convertAlterTableDropConstraint(
+ObjectIdentifier tableIdentifier,
+CatalogTable catalogTable,
+SqlAlterTableDropConstraint alterTableDropConstraint) {
+boolean isPrimaryKey = alterTableDropConstraint.isPrimaryKey();
+Optional oriPrimaryKey =
+catalogTable.getUnresolvedSchema().getPrimaryKey();
+// validate primary key is exists in table

Review Comment:
   nit: 
   ```suggestion
   // validate primary key exists in table
   ```



##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/utils/OperationConverterUtils.java:
##
@@ -117,6 +128,193 @@ public static Operation convertAddReplaceColumns(
 catalogTable.getComment()));
 }
 
+public static Operation convertAlterTableDropConstraint(
+ObjectIdentifier tableIdentifier,
+CatalogTable catalogTable,
+SqlAlterTableDropConstraint alterTableDropConstraint) {
+boolean isPrimaryKey = alterTableDropConstraint.isPrimaryKey();
+Optional oriPrimaryKey =
+catalogTable.getUnresolvedSchema().getPrimaryKey();
+// validate primary key is exists in table
+if (!oriPrimaryKey.isPresent()) {
+throw new ValidationException(
+String.format("Table %s does not exist primary key.", 
tableIdentifier));
+}
+
+String constraintName = null;
+if (alterTableDropConstraint.getConstraintName().isPresent()) {
+constraintName = 
alterTableDropConstraint.getConstraintName().get().getSimple();
+}
+if (!StringUtils.isNullOrWhitespaceOnly(constraintName)
+&& 
!oriPrimaryKey.get().getConstraintName().equals(constraintName)) {
+throw new ValidationException(
+String.format(
+"CONSTRAINT [%s] does not exist in table %s",
+constraintName, tableIdentifier));
+}
+
+return new AlterTableDropConstraintOperation(tableIdentifier, 
isPrimaryKey, constraintName);
+}
+
+public static Operation convertDropWatermark(
+ObjectIdentifier tableIdentifier, CatalogTable catalogTable) {
+

[jira] [Commented] (FLINK-27261) Disable web.cancel.enable for session jobs

2022-04-15 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522797#comment-17522797
 ] 

Gyula Fora commented on FLINK-27261:


But isn’t this something only the session cluster could disable?

> Disable web.cancel.enable for session jobs
> --
>
> Key: FLINK-27261
> URL: https://issues.apache.org/jira/browse/FLINK-27261
> Project: Flink
>  Issue Type: Sub-task
>  Components: Kubernetes Operator
>Reporter: Yang Wang
>Priority: Major
>  Labels: starter
>
> In FLINK-27154, we disable {{web.cancel.enable}} for application cluster. We 
> should also do this for session jobs.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-27264) Add ITCase for concurrent batch overwrite and streaming insert

2022-04-15 Thread Jane Chan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jane Chan updated FLINK-27264:
--
Description: 
!image-2022-04-15-19-26-09-649.png|width=609,height=241!

Add it case for user story

> Add ITCase for concurrent batch overwrite and streaming insert
> --
>
> Key: FLINK-27264
> URL: https://issues.apache.org/jira/browse/FLINK-27264
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table Store
>Affects Versions: table-store-0.1.0
>Reporter: Jane Chan
>Priority: Major
> Fix For: table-store-0.1.0
>
> Attachments: image-2022-04-15-19-26-09-649.png
>
>
> !image-2022-04-15-19-26-09-649.png|width=609,height=241!
> Add it case for user story



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-27264) Add ITCase for concurrent batch overwrite and streaming insert

2022-04-15 Thread Jane Chan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jane Chan updated FLINK-27264:
--
Attachment: image-2022-04-15-19-26-09-649.png

> Add ITCase for concurrent batch overwrite and streaming insert
> --
>
> Key: FLINK-27264
> URL: https://issues.apache.org/jira/browse/FLINK-27264
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table Store
>Affects Versions: table-store-0.1.0
>Reporter: Jane Chan
>Priority: Major
> Fix For: table-store-0.1.0
>
> Attachments: image-2022-04-15-19-26-09-649.png
>
>




--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Created] (FLINK-27264) Add ITCase for concurrent batch overwrite and streaming insert

2022-04-15 Thread Jane Chan (Jira)
Jane Chan created FLINK-27264:
-

 Summary: Add ITCase for concurrent batch overwrite and streaming 
insert
 Key: FLINK-27264
 URL: https://issues.apache.org/jira/browse/FLINK-27264
 Project: Flink
  Issue Type: Sub-task
  Components: Table Store
Affects Versions: table-store-0.1.0
Reporter: Jane Chan
 Fix For: table-store-0.1.0






--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-27237) Partitioned table statement enhancement

2022-04-15 Thread dalongliu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27237?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dalongliu updated FLINK-27237:
--
Description: 
This is an umbrella issue which is used to track the syntax enhancement about 
partitioned table in FLIP [1]. These new syntaxes are very useful for 
partitioned tables, especially for batch job.

The supported statement about partitioned table as follows:
{code:sql}
-- add partition
ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
[PARTITION  ...] } 

-- drop partition
ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
[PARTITION  ...] } [PURGE]

-- rename partition
ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
;

-- show partitions
SHOW PARTITIONS table_name [PARTITION ]


:: 
  (partition_col_name=partition_col_val, ...)
 {code}
 

Reference:

[1]: 
[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support]

[[2] 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]

[3]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]

[4]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]

  was:
This is an umbrella issue which is used to track the syntax enhancement about 
partitioned table. These new syntaxes are very useful for partitioned tables, 
especially for batch job.

The supported statement about partitioned table as follows:
{code:sql}
-- add partition
ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
[PARTITION  ...] } 

-- drop partition
ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
[PARTITION  ...] } [PURGE]

-- rename partition
ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
;

-- show partitions
SHOW PARTITIONS table_name [PARTITION ]


:: 
  (partition_col_name=partition_col_val, ...)
 {code}
 

Reference:

[1]: 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support

[[2] 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]

[3]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]

[4]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]


> Partitioned table statement enhancement
> ---
>
> Key: FLINK-27237
> URL: https://issues.apache.org/jira/browse/FLINK-27237
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: dalongliu
>Priority: Major
> Fix For: 1.16.0
>
>
> This is an umbrella issue which is used to track the syntax enhancement about 
> partitioned table in FLIP [1]. These new syntaxes are very useful for 
> partitioned tables, especially for batch job.
> The supported statement about partitioned table as follows:
> {code:sql}
> -- add partition
> ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
> [PARTITION  ...] } 
> -- drop partition
> ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
> [PARTITION  ...] } [PURGE]
> -- rename partition
> ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
> ;
> -- show partitions
> SHOW PARTITIONS table_name [PARTITION ]
> :: 
>   (partition_col_name=partition_col_val, ...)
>  {code}
>  
> Reference:
> [1]: 
> [https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support]
> [[2] 
> https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]
> [3]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]
> [4]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] zoltar9264 commented on pull request #19252: [FLINK-25867][docs] translate ChangelogBackend documentation to chinese

2022-04-15 Thread GitBox


zoltar9264 commented on PR #19252:
URL: https://github.com/apache/flink/pull/19252#issuecomment-1100046811

   Thank you very much @Myasuka , your suggestion does look better, learn a lot 
!
   I hive committed those change and squash them into first commit in this pr. 
(Last three commits are hotfix for other issue.)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] flinkbot commented on pull request #19490: [FLINK-27263][table] Rename the metadata column to the user specified…

2022-04-15 Thread GitBox


flinkbot commented on PR #19490:
URL: https://github.com/apache/flink/pull/19490#issuecomment-1100045930

   
   ## CI report:
   
   * b98759b38d05a732ef3c9ceabfc844dfaf4bd2e8 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-27018) timestamp missing end zero when outputing to kafka

2022-04-15 Thread jeff-zou (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522793#comment-17522793
 ] 

jeff-zou commented on FLINK-27018:
--

I think it's better to be consistent  in converting, because many scenarios 
require equals, For example,  After I sinked the result to ES, I need to use 
the original value to find out if the result exsits.

> timestamp missing end  zero when outputing to kafka
> ---
>
> Key: FLINK-27018
> URL: https://issues.apache.org/jira/browse/FLINK-27018
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.13.5
>Reporter: jeff-zou
>Priority: Major
> Attachments: kafka.png
>
>
> the bug is described as follows:
>  
> {code:java}
> data in source:
>  2022-04-02 03:34:21.260
> but after sink by sql, data in kafka:
>  2022-04-02 03:34:21.26
> {code}
>  
> data miss end zero in kafka.
>  
> sql:
> {code:java}
> create kafka_table(stime stimestamp) with ('connector'='kafka','format' = 
> 'json');
> insert into kafka_table select stime from (values(timestamp '2022-04-02 
> 03:34:21.260')){code}
> the value in kafka is : \{"stime":"2022-04-02 03:34:21.26"}, missed end zero.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-27263) Rename the metadata column to the user specified name in DDL

2022-04-15 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27263?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-27263:
---
Labels: pull-request-available  (was: )

> Rename the metadata column to the user specified name in DDL
> 
>
> Key: FLINK-27263
> URL: https://issues.apache.org/jira/browse/FLINK-27263
> Project: Flink
>  Issue Type: Improvement
>Reporter: Shengkai Fang
>Priority: Major
>  Labels: pull-request-available
>
> Currently the source output the metadata column with the "$metadata"+ 
> metadata key as the column name. It's better that keep the name align with 
> the user name specified in the DDL,e.g. use event_time as the name if the 
> user specifies`event_time` TIMESTAMP(3) FROM 'timestamp'.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] fsk119 opened a new pull request, #19490: [FLINK-27263][table] Rename the metadata column to the user specified…

2022-04-15 Thread GitBox


fsk119 opened a new pull request, #19490:
URL: https://github.com/apache/flink/pull/19490

   … name in DDL
   
   
   
   ## What is the purpose of the change
   
   *Currently the source output the metadata column with the `"$metadata"+ 
metadata key` as the column name. It's better that keep the name align with the 
user name specified in the DDL,e.g. use `event_time` as the name if the user 
specifies `event_time TIMESTAMP(3) FROM 'timestamp'`.*
   
   
   ## Brief change log
   
 - *Make call in the projection to do the rename*
   
   
   ## Verifying this change
   
   
   This change is already covered by existing tests in the 
`PushProjectIntoTableSourceScanRuleTest`.
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / **no** / don't 
know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ **not documented**)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-27263) Rename the metadata column to the user specified name in DDL

2022-04-15 Thread Shengkai Fang (Jira)
Shengkai Fang created FLINK-27263:
-

 Summary: Rename the metadata column to the user specified name in 
DDL
 Key: FLINK-27263
 URL: https://issues.apache.org/jira/browse/FLINK-27263
 Project: Flink
  Issue Type: Improvement
Reporter: Shengkai Fang


Currently the source output the metadata column with the "$metadata"+ metadata 
key as the column name. It's better that keep the name align with the user name 
specified in the DDL,e.g. use event_time as the name if the user 
specifies`event_time` TIMESTAMP(3) FROM 'timestamp'.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] MartijnVisser merged pull request #19401: [FLINK-25716][docs-zh] Translate "Streaming Concepts" page of "Applic…

2022-04-15 Thread GitBox


MartijnVisser merged PR #19401:
URL: https://github.com/apache/flink/pull/19401


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] MartijnVisser commented on pull request #19401: [FLINK-25716][docs-zh] Translate "Streaming Concepts" page of "Applic…

2022-04-15 Thread GitBox


MartijnVisser commented on PR #19401:
URL: https://github.com/apache/flink/pull/19401#issuecomment-1100038830

   @snailHumming Thanks for the PR. I've squashed and rebased commits. The 
successful CI run was in 
https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=34715.
 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] MartijnVisser commented on pull request #19489: in filesystem.md, flink comes with four built-in BulkWriter factories…

2022-04-15 Thread GitBox


MartijnVisser commented on PR #19489:
URL: https://github.com/apache/flink/pull/19489#issuecomment-1100037906

   @chenlei677 Thanks for the PR. Please check and follow the code contribution 
guide at https://flink.apache.org/contributing/contribute-code.html and update 
your PR, else it can't get merged.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] MartijnVisser commented on pull request #19487: in filesystem.md, the format parameter described is incomplete.

2022-04-15 Thread GitBox


MartijnVisser commented on PR #19487:
URL: https://github.com/apache/flink/pull/19487#issuecomment-1100037548

   @chenlei677 Thanks for the PR. However, the list of formats in the 
documentation is not an extensive list of all formats. I don't think adding one 
adds a lot of value in this case.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] MartijnVisser closed pull request #19485: Mssql branch

2022-04-15 Thread GitBox


MartijnVisser closed pull request #19485: Mssql branch
URL: https://github.com/apache/flink/pull/19485


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] MartijnVisser commented on pull request #19485: Mssql branch

2022-04-15 Thread GitBox


MartijnVisser commented on PR #19485:
URL: https://github.com/apache/flink/pull/19485#issuecomment-1100036893

   @yinweiwen Please check and follow the code contribution guide at 
https://flink.apache.org/contributing/contribute-code.html - This PR is 
incorrect, therefore I'm closing it


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-22246) when use HiveCatalog create table , can't set Table owner property correctly

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22246:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> when use HiveCatalog create table ,  can't set Table owner property correctly
> -
>
> Key: FLINK-22246
> URL: https://issues.apache.org/jira/browse/FLINK-22246
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1, 1.12.0
>Reporter: xiangtao
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> when i use HiveCatalog create table in sql-client ,  i found it can't set 
> Hive Table `owner` property correctly.  i debug code , i found  in 
> `HiveCatalog.createTable` method 
> {code:java}
> Table hiveTable =
>  org.apache.hadoop.hive.ql.metadata.Table.getEmptyTable(
>  tablePath.getDatabaseName(), tablePath.getObjectName());
> {code}
>  
> this get hiveTable  obj , owner field is null . beacuse it set owner through 
> {code:java}
> t.setOwner(SessionState.getUserFromAuthenticator()); 
> {code}
>  
> but SessionState is null .  
> Fix this bug , we can add one code in HiveCatalog.open method . 
> {code:java}
> SessionState.setCurrentSessionState(new SessionState(hiveConf));
> {code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-21699) SecurityUtils JaasModule make JVM dirty

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-21699:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> SecurityUtils JaasModule make JVM dirty
> ---
>
> Key: FLINK-21699
> URL: https://issues.apache.org/jira/browse/FLINK-21699
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Configuration
>Affects Versions: 1.12.2
>Reporter: tonychan
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
> Attachments: image-2021-03-10-13-19-04-251.png
>
>
> System.setProperty(JAVA_SECURITY_AUTH_LOGIN_CONFIG, 
> configFile.getAbsolutePath());
> "java.security.auth.login.config": "/tmp/jaas-4611020726588581730.conf",
> "zookeeper.sasl.client": "true",
>  
> !image-2021-03-10-13-19-04-251.png!



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-22644) Translate "Native Kubernetes" page into Chinese

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22644:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor auto-unassigned 
pull-request-available  (was: auto-deprioritized-major auto-unassigned 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Translate "Native Kubernetes" page into Chinese
> ---
>
> Key: FLINK-22644
> URL: https://issues.apache.org/jira/browse/FLINK-22644
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Yuchen Cheng
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned, pull-request-available
>
> The page url is 
> https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/deployment/resource-providers/native_kubernetes/



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-22663) Release YARN resource very slow when cancel the job after some NodeManagers shutdown

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22663?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22663:
---
  Labels: YARN auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: YARN auto-deprioritized-major 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Release YARN resource very slow when cancel the job after some NodeManagers 
> shutdown
> 
>
> Key: FLINK-22663
> URL: https://issues.apache.org/jira/browse/FLINK-22663
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.12.2
>Reporter: Jinhong Liu
>Priority: Not a Priority
>  Labels: YARN, auto-deprioritized-major, 
> auto-deprioritized-minor, pull-request-available
>
> When I test flink on YARN, there is a case that may cause some problems.
> Hadoop Version: 2.7.3
> Flink Version: 1.12.2
> I deploy a flink job on YARN, when the job is running I stop one NodeManager, 
> after one or two minutes, the job is auto recovered. But in this situation, 
> if I cancel the job, the containers cannot be released immediately, there are 
> still some containers that are running include the app master. About 5 
> minutes later, these containers exit, and about 10 minutes later the app 
> master exit.
> I check the log of app master, seems it try to stop the containers on the 
> NodeManger which I have already stopped.
> {code:java}
> 2021-05-14 06:15:17,389 INFO  
> org.apache.flink.runtime.executiongraph.ExecutionGraph   [] - Job class 
> tv.freewheel.reporting.fastlane.Fastlane$ (da883ab39a7a82e4d45a3803bc77dd6f) 
> switched from state CANCELLING to CANCELED.
> 2021-05-14 06:15:17,389 INFO  
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator[] - Stopping 
> checkpoint coordinator for job da883ab39a7a82e4d45a3803bc77dd6f.
> 2021-05-14 06:15:17,390 INFO  
> org.apache.flink.runtime.checkpoint.StandaloneCompletedCheckpointStore [] - 
> Shutting down
> 2021-05-14 06:15:17,408 INFO  
> org.apache.flink.runtime.dispatcher.MiniDispatcher   [] - Job 
> da883ab39a7a82e4d45a3803bc77dd6f reached globally terminal state CANCELED.
> 2021-05-14 06:15:17,409 INFO  
> org.apache.flink.runtime.dispatcher.MiniDispatcher   [] - Shutting 
> down cluster with state CANCELED, jobCancelled: true, executionMode: DETACHED
> 2021-05-14 06:15:17,409 INFO  
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint[] - Shutting 
> YarnJobClusterEntrypoint down with application status CANCELED. Diagnostics 
> null.
> 2021-05-14 06:15:17,409 INFO  
> org.apache.flink.runtime.jobmaster.MiniDispatcherRestEndpoint [] - Shutting 
> down rest endpoint.
> 2021-05-14 06:15:17,420 INFO  org.apache.flink.runtime.jobmaster.JobMaster
>  [] - Stopping the JobMaster for job class 
> tv.freewheel.reporting.fastlane.Fastlane$(da883ab39a7a82e4d45a3803bc77dd6f).
> 2021-05-14 06:15:17,422 INFO  
> org.apache.flink.runtime.jobmaster.MiniDispatcherRestEndpoint [] - Removing 
> cache directory 
> /tmp/flink-web-af72a00c-0ddd-4e5e-a62c-8244d6caa552/flink-web-ui
> 2021-05-14 06:15:17,432 INFO  
> org.apache.flink.runtime.jobmaster.MiniDispatcherRestEndpoint [] - 
> http://ip-10-23-19-197.ec2.internal:43811 lost leadership
> 2021-05-14 06:15:17,432 INFO  
> org.apache.flink.runtime.jobmaster.MiniDispatcherRestEndpoint [] - Shut down 
> complete.
> 2021-05-14 06:15:17,436 INFO  
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
> Shut down cluster because application is in CANCELED, diagnostics null.
> 2021-05-14 06:15:17,436 INFO  org.apache.flink.yarn.YarnResourceManagerDriver 
>  [] - Unregister application from the YARN Resource Manager with 
> final status KILLED.
> 2021-05-14 06:15:17,458 INFO  
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl [] - Suspending 
> SlotPool.
> 2021-05-14 06:15:17,458 INFO  org.apache.flink.runtime.jobmaster.JobMaster
>  [] - Close ResourceManager connection 
> 493862ba148679a4f16f7de5ffaef665: Stopping JobMaster for job class 
> tv.freewheel.reporting.fastlane.Fastlane$(da883ab39a7a82e4d45a3803bc77dd6f)..
> 2021-05-14 06:15:17,458 INFO  
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl [] - Stopping 
> SlotPool.
> 2021-05-14 06:15:17,482 INFO  
> org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl[] - Waiting for 
> application to be successfully unregistered.
> 2021-05-14 06:15:17,566 INFO  

[jira] [Updated] (FLINK-22437) Miss adding parallesim for filter operator in batch mode

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22437:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor auto-unassigned 
pull-request-available  (was: auto-deprioritized-major auto-unassigned 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Miss adding parallesim for filter operator in batch mode
> 
>
> Key: FLINK-22437
> URL: https://issues.apache.org/jira/browse/FLINK-22437
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.12.2
>Reporter: zoucao
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned, pull-request-available
>
> when I execute batch sql as follow in flink-1.12.2, I found lots of small 
> files in hdfs. In filesystem connector, `GroupedPartitionWriter` will be 
> used, and it close the last partiton if a new record does not belong to the 
> existing partition. The phenomenon occurred if there are more than one 
> partiton's records are sent to filesystem sink at the same time. Hive source 
> can determine parallesim by the number of file and partition, and the 
> parallesim will extended by sort operator,  but in 
> `CommonPhysicalSink#createSinkTransformation`,a filter operator will be add 
> to support `SinkNotNullEnforcer`, there is no parallesim set for it, so 
> filesystem sink operator can not get the correct parallesim from inputstream.
> {code:java}
> CREATE CATALOG myHive with (
> 'type'='hive',
> 'property-version'='1',
> 'default-database' = 'flink_sql_online_test'
> );
> -- SET table.sql-dialect=hive;
> -- CREATE TABLE IF NOT EXISTS myHive.flink_sql_online_test.hive_sink (
> --`timestamp` BIGINT,
> --`time` STRING,
> --id BIGINT,
> --product STRING,
> --price DOUBLE,
> --canSell STRING,
> --selledNum BIGINT
> -- ) PARTITIONED BY (
> --dt STRING,
> --`hour` STRING,
> --   `min` STRING
> -- ) TBLPROPERTIES (
> --'partition.time-extractor.timestamp-pattern'='$dt $hr:$min:00',
> --'sink.partition-commit.trigger'='partition-time',
> --'sink.partition-commit.delay'='1 min',
> --'sink.partition-commit.policy.kind'='metastore,success-file'
> -- );
> create table fs_sink (
> `timestamp` BIGINT,
> `time` STRING,
> id BIGINT,
> product STRING,
> price DOUBLE,
> canSell STRING,
> selledNum BIGINT,
> dt STRING,
> `hour` STRING,
> `min` STRING
> ) PARTITIONED BY (dt, `hour`, `min`) with (
> 'connector'='filesystem',
> 'path'='hdfs://',
> 'format'='csv'
> );
> insert into fs_sink
> select * from myHive.flink_sql_online_test.hive_sink;
> {code}
> I think this problem can be fixed by adding a parallesim for it just like
> {code:java}
> val dataStream = new DataStream(env, inputTransformation).filter(enforcer)
>   .setParallelism(inputTransformation.getParallelism)
> {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-22679) code optimization:Transformation.equals

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22679:
---
  Labels: auto-deprioritized-minor pull-request-available  (was: 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


>  code optimization:Transformation.equals
> 
>
> Key: FLINK-22679
> URL: https://issues.apache.org/jira/browse/FLINK-22679
> Project: Flink
>  Issue Type: Improvement
>  Components: API / DataStream
>Affects Versions: 1.13.0
>Reporter: huzeming
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, pull-request-available
>
> code optimization:Transformation.equals , line : 550
> {code:java}
> // old
> return outputType != null ? outputType.equals(that.outputType) : 
> that.outputType == null;
> // new 
> return Objects.equals(outputType, that.outputType);{code}
> I think after change  it will be more readable



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-22798) Exception handling of transactional producer

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22798?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22798:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Exception handling of transactional producer
> 
>
> Key: FLINK-22798
> URL: https://issues.apache.org/jira/browse/FLINK-22798
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Reporter: Jin Xu
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> [https://github.com/apache/flink/blob/8e90425462f287b94e867ba778d75abb193763bd/flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer.java#L1013]
>  
> [https://github.com/apache/flink/blob/8e90425462f287b94e867ba778d75abb193763bd/flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer.java#L1047]
>  
> current implementation of commit and abort do not handle critical exception 
> and try to recycle/reuse the producer.
>  
> Kafka doc 
> [https://kafka.apache.org/11/javadoc/index.html?org/apache/kafka/clients/producer/KafkaProducer.html]
> recommends catch critical exceptions and should not recycle the producer.
> {code:java}
> catch (ProducerFencedException | OutOfOrderSequenceException | 
> AuthorizationException e) 
> {code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-22526) Rename TopicId to topic in Kafka related code

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-22526:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Rename TopicId to topic in Kafka related code
> -
>
> Key: FLINK-22526
> URL: https://issues.apache.org/jira/browse/FLINK-22526
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: dengziming
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> TopicId is a new concept introduced in Kafka 2.8.0, we can use it to 
> produce/consume in future Kafka version, so we should avoid use topicId when 
> referring to a topic.
> for more details: 
> https://cwiki.apache.org/confluence/display/KAFKA/KIP-516:+Topic+Identifiers



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-23158) Source transformation is not added to the StreamExecutionEnvironment explicitly

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-23158:
---
  Labels: auto-deprioritized-major auto-deprioritized-minor 
pull-request-available  (was: auto-deprioritized-major pull-request-available 
stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Source transformation is not added to the StreamExecutionEnvironment 
> explicitly
> ---
>
> Key: FLINK-23158
> URL: https://issues.apache.org/jira/browse/FLINK-23158
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.14.0
>Reporter: Yun Gao
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> pull-request-available
>
> Currently for the implementation of `StreamExecutionEnvironment#fromSource()` 
> and `StreamExecutionEnvironment#addSource()`, the SourceTransformation is not 
> added to the transformation list explicitly, this make the job with a single 
> source could not run directly. For example, 
> {code:java}
> StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>   env.addSource(new SourceFunction() {
>   @Override
>   public void run(SourceContext sourceContext) 
> throws Exception {
>   }
>   @Override
>   public void cancel() {
>   }
>   });
>   env.execute();
> {code}
> would throws the exception:
> {code:java}
> Exception in thread "main" java.lang.IllegalStateException: No operators 
> defined in streaming topology. Cannot execute.
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraphGenerator(StreamExecutionEnvironment.java:2019)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2010)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:1995)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1834)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1817)
>   at test.SingleSourceTest.main(SingleSourceTest.java:41)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Updated] (FLINK-23163) Add debug ability for kubernetes client

2022-04-15 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23163?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-23163:
---
  Labels: auto-deprioritized-minor pull-request-available  (was: 
pull-request-available stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Add debug ability for kubernetes client 
> 
>
> Key: FLINK-23163
> URL: https://issues.apache.org/jira/browse/FLINK-23163
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / Kubernetes
>Affects Versions: 1.13.0
>Reporter: Aitozi
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, pull-request-available
>
> add option to enable http client debug log



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] flinkbot commented on pull request #19489: in filesystem.md, flink comes with four built-in BulkWriter factories…

2022-04-15 Thread GitBox


flinkbot commented on PR #19489:
URL: https://github.com/apache/flink/pull/19489#issuecomment-1100028822

   
   ## CI report:
   
   * ab93146354ccd9d981e5fb854dcdfb9ce33047e1 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-27218) Serializer in OperatorState has not been updated when new Serializers are NOT incompatible

2022-04-15 Thread Yue Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27218?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522779#comment-17522779
 ] 

Yue Ma commented on FLINK-27218:


[~yunta] ok i'm glad to do this fix, i'll submit a pr soon

> Serializer in OperatorState has not been updated when new Serializers are NOT 
> incompatible
> --
>
> Key: FLINK-27218
> URL: https://issues.apache.org/jira/browse/FLINK-27218
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / State Backends
>Affects Versions: 1.15.1
>Reporter: Yue Ma
>Priority: Major
> Attachments: image-2022-04-13-14-50-10-921.png
>
>
> OperatorState such as *BroadcastState* or *PartitionableListState*  can only 
> be constructed via {*}DefaultOperatorStateBackend{*}. But when 
> *BroadcastState* or *PartitionableListState* Serializer changes after we 
> restart the job , it seems to have the following problems .
> As an example, we can see how PartitionableListState is initialized.
> First, RestoreOperation will construct a restored PartitionableListState 
> based on the information in the snapshot.
> Then StateMetaInfo in partitionableListState will be updated  as the 
> following code
> {code:java}
> TypeSerializerSchemaCompatibility stateCompatibility =
>                 
> restoredPartitionableListStateMetaInfo.updatePartitionStateSerializer(newPartitionStateSerializer);
> partitionableListState.setStateMetaInfo(restoredPartitionableListStateMetaInfo);{code}
> The main problem is that there is also an *internalListCopySerializer* in 
> *PartitionableListState* that is built using the previous Serializer and it 
> has not been updated. 
> Therefore, when we update the StateMetaInfo, the *internalListCopySerializer* 
> also needs to be updated.
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] chenlei677 opened a new pull request, #19489: in filesystem.md, flink comes with four built-in BulkWriter factories…

2022-04-15 Thread GitBox


chenlei677 opened a new pull request, #19489:
URL: https://github.com/apache/flink/pull/19489

   ## What is the purpose of the change
   in filesystem.md, flink comes with four built-in BulkWriter factories, in 
fact, the list has five.
   
   
   ## Brief change log
   
   in filesystem.md, flink comes with four built-in BulkWriter factories, in 
fact, the list has five.
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   yes
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature?  no


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] Myasuka commented on a diff in pull request #19252: [FLINK-25867][docs] translate ChangelogBackend documentation to chinese

2022-04-15 Thread GitBox


Myasuka commented on code in PR #19252:
URL: https://github.com/apache/flink/pull/19252#discussion_r851163050


##
docs/content.zh/docs/ops/state/state_backends.md:
##
@@ -306,77 +329,75 @@ public class MyOptionsFactory implements 
ConfigurableRocksDBOptionsFactory {
 
 {{< top >}}
 
-## Enabling Changelog
+
 
-{{< hint warning >}} This feature is in experimental status. {{< /hint >}}
+## 开启 Changelog
 
-{{< hint warning >}} Enabling Changelog may have a negative performance impact 
on your application (see below). {{< /hint >}}
+{{< hint warning >}} 该功能处于实验状态。 {{< /hint >}}
 
-### Introduction
+{{< hint warning >}} 开启 Changelog 可能会给您的应用带来性能损失。(见下文) {{< /hint >}}
 
-Changelog is a feature that aims to decrease checkpointing time and, 
therefore, end-to-end latency in exactly-once mode.
+
 
-Most commonly, checkpoint duration is affected by:
+### 介绍
 
-1. Barrier travel time and alignment, addressed by
-   [Unaligned checkpoints]({{< ref 
"docs/ops/state/checkpointing_under_backpressure#unaligned-checkpoints" >}})
-   and [Buffer debloating]({{< ref 
"docs/ops/state/checkpointing_under_backpressure#buffer-debloating" >}})
-2. Snapshot creation time (so-called synchronous phase), addressed by 
asynchronous snapshots (mentioned [above]({{<
-   ref "#the-embeddedrocksdbstatebackend">}}))
-4. Snapshot upload time (asynchronous phase)
+Changelog 是一项旨在减少 checkpointing 时间的功能,因此也可以减少 exactly-once 模式下的端到端延迟。
 
-Upload time can be decreased by [incremental checkpoints]({{< ref 
"#incremental-checkpoints" >}}).
-However, most incremental state backends perform some form of compaction 
periodically, which results in re-uploading the
-old state in addition to the new changes. In large deployments, the 
probability of at least one task uploading lots of
-data tends to be very high in every checkpoint.
+一般情况下 checkpoint 持续时间受如下因素影响:

Review Comment:
   ```suggestion
   一般情况下 checkpoint 的持续时间受如下因素影响:
   ```



##
docs/content.zh/docs/ops/state/state_backends.md:
##
@@ -306,77 +329,75 @@ public class MyOptionsFactory implements 
ConfigurableRocksDBOptionsFactory {
 
 {{< top >}}
 
-## Enabling Changelog
+
 
-{{< hint warning >}} This feature is in experimental status. {{< /hint >}}
+## 开启 Changelog
 
-{{< hint warning >}} Enabling Changelog may have a negative performance impact 
on your application (see below). {{< /hint >}}
+{{< hint warning >}} 该功能处于实验状态。 {{< /hint >}}
 
-### Introduction
+{{< hint warning >}} 开启 Changelog 可能会给您的应用带来性能损失。(见下文) {{< /hint >}}
 
-Changelog is a feature that aims to decrease checkpointing time and, 
therefore, end-to-end latency in exactly-once mode.
+
 
-Most commonly, checkpoint duration is affected by:
+### 介绍
 
-1. Barrier travel time and alignment, addressed by
-   [Unaligned checkpoints]({{< ref 
"docs/ops/state/checkpointing_under_backpressure#unaligned-checkpoints" >}})
-   and [Buffer debloating]({{< ref 
"docs/ops/state/checkpointing_under_backpressure#buffer-debloating" >}})
-2. Snapshot creation time (so-called synchronous phase), addressed by 
asynchronous snapshots (mentioned [above]({{<
-   ref "#the-embeddedrocksdbstatebackend">}}))
-4. Snapshot upload time (asynchronous phase)
+Changelog 是一项旨在减少 checkpointing 时间的功能,因此也可以减少 exactly-once 模式下的端到端延迟。
 
-Upload time can be decreased by [incremental checkpoints]({{< ref 
"#incremental-checkpoints" >}}).
-However, most incremental state backends perform some form of compaction 
periodically, which results in re-uploading the
-old state in addition to the new changes. In large deployments, the 
probability of at least one task uploading lots of
-data tends to be very high in every checkpoint.
+一般情况下 checkpoint 持续时间受如下因素影响:
 
-With Changelog enabled, Flink uploads state changes continuously and forms a 
changelog. On checkpoint, only the relevant
-part of this changelog needs to be uploaded. The configured state backend is 
snapshotted in the
-background periodically. Upon successful upload, the changelog is truncated.
+1. Barrier 到达和对齐时间,可以通过 [Unaligned checkpoints]({{< ref 
"docs/ops/state/checkpointing_under_backpressure#unaligned-checkpoints" >}}) 和 
[Buffer debloating]({{< ref 
"docs/ops/state/checkpointing_under_backpressure#buffer-debloating" >}}) 解决。
 
-As a result, asynchronous phase duration is reduced, as well as synchronous 
phase - because no data needs to be flushed
-to disk. In particular, long-tail latency is improved.
+2. 快照制作时间(所谓同步阶段), 可以通过异步快照解决(如[上文]({{<
+   ref "#the-embeddedrocksdbstatebackend">}})所述)。
 
-However, resource usage is higher:
+3. 快照上传时间(异步阶段)。
 
-- more files are created on DFS
-- more files can be left undeleted DFS (this will be addressed in the future 
versions in FLINK-25511 and FLINK-25512)
-- more IO bandwidth is used to upload state changes
-- more CPU used to serialize state changes
-- more memory used by Task Managers to buffer state changes
+可以用[增量 checkpoints]({{< ref "#incremental-checkpoints" >}}) 

[GitHub] [flink-kubernetes-operator] wangyang0918 commented on pull request #165: [FLINK-26140] Support rollback strategies

2022-04-15 Thread GitBox


wangyang0918 commented on PR #165:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/165#issuecomment-1100014552

   @gyfora No hurry, take your time.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-27215) JDBC sink transiently deleted a record because of -u message of that record

2022-04-15 Thread Shengkai Fang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522764#comment-17522764
 ] 

Shengkai Fang commented on FLINK-27215:
---

Hi, [~yulei0824]. 

It's up to the planner to determine whether the upstream operator sends the -U. 
The logic just about  how the jdbc sink processes the -U messages. If you are 
interested about the ChangelogMode inference, you can take a look at the 
FlinkChangelogModeInferenceProgram. I think we have already made the 
optimizaiton.

> JDBC sink transiently deleted a record because of -u message of that record
> ---
>
> Key: FLINK-27215
> URL: https://issues.apache.org/jira/browse/FLINK-27215
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.12.7, 1.13.5, 1.14.3
>Reporter: tim yu
>Priority: Major
>
> A record is deleted transiently when using JDBC sink in upsert mode.
> The -U message is processed as delete operation in class 
> TableBufferReducedStatementExecutor.
> The following codes show how to process -U message:
> {code:java}
> /**
>  * Returns true if the row kind is INSERT or UPDATE_AFTER, returns false 
> if the row kind is
>  * DELETE or UPDATE_BEFORE.
>  */
> private boolean changeFlag(RowKind rowKind) {
> switch (rowKind) {
> case INSERT:
> case UPDATE_AFTER:
> return true;
> case DELETE:
> case UPDATE_BEFORE:
> return false;
> default:
> throw new UnsupportedOperationException(
> String.format(
> "Unknown row kind, the supported row kinds 
> is: INSERT, UPDATE_BEFORE, UPDATE_AFTER,"
> + " DELETE, but get: %s.",
> rowKind));
> }
> }
> @Override
> public void executeBatch() throws SQLException {
> for (Map.Entry> entry : 
> reduceBuffer.entrySet()) {
> if (entry.getValue().f0) {
> upsertExecutor.addToBatch(entry.getValue().f1);
> } else {
> // delete by key
> deleteExecutor.addToBatch(entry.getKey());
> }
> }
> upsertExecutor.executeBatch();
> deleteExecutor.executeBatch();
> reduceBuffer.clear();
> }
> {code}
> If -U and +U messages of one record are executed separately in different JDBC 
> batches, that record will be deleted transiently in external database and 
> then insert a new updated record to it. In fact, this record should be merely 
> updated once in the external database.
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] luoyuxia commented on a diff in pull request #19424: [FLINK-22732][table] Restrict ALTER TABLE from setting empty table op…

2022-04-15 Thread GitBox


luoyuxia commented on code in PR #19424:
URL: https://github.com/apache/flink/pull/19424#discussion_r851182447


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/operations/SqlToOperationConverter.java:
##
@@ -547,6 +547,12 @@ private Operation convertAlterTableOptions(
 ObjectIdentifier tableIdentifier,
 CatalogTable oldTable,
 SqlAlterTableOptions alterTableOptions) {
+Map setOptions =

Review Comment:
   Will it be better to do the validation when parse `alter table set` in 
`parserImpls.ftl`
   For me,  it looks like a restriction in sql synax level.  From my sense, 
`alter table cat1.db1.tb1 set ()` is like a illegal sql.
   Anyway, looks good to me for  either way .



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-27237) Partitioned table statement enhancement

2022-04-15 Thread dalongliu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27237?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dalongliu updated FLINK-27237:
--
Description: 
This is an umbrella issue which is used to track the syntax enhancement about 
partitioned table. These new syntaxes are very useful for partitioned tables, 
especially for batch job.

The supported statement about partitioned table as follows:
{code:sql}
-- add partition
ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
[PARTITION  ...] } 

-- drop partition
ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
[PARTITION  ...] } [PURGE]

-- rename partition
ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
;

-- show partitions
SHOW PARTITIONS table_name [PARTITION ]


:: 
  (partition_col_name=partition_col_val, ...)
 {code}
 

Reference:

[1]: 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support

[[2] 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]

[3]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]

[4]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]

  was:
This is an umbrella issue which is used to discuss the syntax enhancement about 
partitioned table. These new syntaxes are very useful for partitioned tables, 
especially for batch job.

Therefore, I propose to support the following statement about partitioned table:
{code:sql}
-- add partition
ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
[PARTITION  ...] } 

-- drop partition
ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
[PARTITION  ...] } [PURGE]

-- rename partition
ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
;

-- show partitions
SHOW PARTITIONS table_name [PARTITION ]


:: 
  (partition_col_name=partition_col_val, ...)
 {code}
 

Reference:

[1]: 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]

[2]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]

[3]: 
[https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]


> Partitioned table statement enhancement
> ---
>
> Key: FLINK-27237
> URL: https://issues.apache.org/jira/browse/FLINK-27237
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: dalongliu
>Priority: Major
> Fix For: 1.16.0
>
>
> This is an umbrella issue which is used to track the syntax enhancement about 
> partitioned table. These new syntaxes are very useful for partitioned tables, 
> especially for batch job.
> The supported statement about partitioned table as follows:
> {code:sql}
> -- add partition
> ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
> [PARTITION  ...] } 
> -- drop partition
> ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
> [PARTITION  ...] } [PURGE]
> -- rename partition
> ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
> ;
> -- show partitions
> SHOW PARTITIONS table_name [PARTITION ]
> :: 
>   (partition_col_name=partition_col_val, ...)
>  {code}
>  
> Reference:
> [1]: 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support
> [[2] 
> https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]
> [3]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]
> [4]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Comment Edited] (FLINK-27237) Partitioned table statement enhancement

2022-04-15 Thread dalongliu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27237?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522755#comment-17522755
 ] 

dalongliu edited comment on FLINK-27237 at 4/15/22 9:31 AM:


We have discussed and voted this ticket in FLIP-63 
[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support],so
 it is the continuation of FLIP-63.


was (Author: lsy):
We have discussed and voted this ticket in 
[FLIP-63|[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support]]
 ,so it is the continuation of FLIP-63.

> Partitioned table statement enhancement
> ---
>
> Key: FLINK-27237
> URL: https://issues.apache.org/jira/browse/FLINK-27237
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: dalongliu
>Priority: Major
> Fix For: 1.16.0
>
>
> This is an umbrella issue which is used to discuss the syntax enhancement 
> about partitioned table. These new syntaxes are very useful for partitioned 
> tables, especially for batch job.
> Therefore, I propose to support the following statement about partitioned 
> table:
> {code:sql}
> -- add partition
> ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
> [PARTITION  ...] } 
> -- drop partition
> ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
> [PARTITION  ...] } [PURGE]
> -- rename partition
> ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
> ;
> -- show partitions
> SHOW PARTITIONS table_name [PARTITION ]
> :: 
>   (partition_col_name=partition_col_val, ...)
>  {code}
>  
> Reference:
> [1]: 
> [https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]
> [2]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]
> [3]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Comment Edited] (FLINK-27237) Partitioned table statement enhancement

2022-04-15 Thread dalongliu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27237?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522755#comment-17522755
 ] 

dalongliu edited comment on FLINK-27237 at 4/15/22 9:30 AM:


We have discussed and voted this ticket in 
[FLIP-63|[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support]]
 ,so it is the continuation of FLIP-63.


was (Author: lsy):
We have discussed and voted this ticket in 
[FLIP-63|[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support]]
 So it is the continuation of FLIP-63.

> Partitioned table statement enhancement
> ---
>
> Key: FLINK-27237
> URL: https://issues.apache.org/jira/browse/FLINK-27237
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: dalongliu
>Priority: Major
> Fix For: 1.16.0
>
>
> This is an umbrella issue which is used to discuss the syntax enhancement 
> about partitioned table. These new syntaxes are very useful for partitioned 
> tables, especially for batch job.
> Therefore, I propose to support the following statement about partitioned 
> table:
> {code:sql}
> -- add partition
> ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
> [PARTITION  ...] } 
> -- drop partition
> ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
> [PARTITION  ...] } [PURGE]
> -- rename partition
> ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
> ;
> -- show partitions
> SHOW PARTITIONS table_name [PARTITION ]
> :: 
>   (partition_col_name=partition_col_val, ...)
>  {code}
>  
> Reference:
> [1]: 
> [https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]
> [2]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]
> [3]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Commented] (FLINK-27261) Disable web.cancel.enable for session jobs

2022-04-15 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522757#comment-17522757
 ] 

Yang Wang commented on FLINK-27261:
---

Hmm. I mean to disable for all the session jobs, not a per job level.

> Disable web.cancel.enable for session jobs
> --
>
> Key: FLINK-27261
> URL: https://issues.apache.org/jira/browse/FLINK-27261
> Project: Flink
>  Issue Type: Sub-task
>  Components: Kubernetes Operator
>Reporter: Yang Wang
>Priority: Major
>  Labels: starter
>
> In FLINK-27154, we disable {{web.cancel.enable}} for application cluster. We 
> should also do this for session jobs.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Commented] (FLINK-27237) Partitioned table statement enhancement

2022-04-15 Thread dalongliu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27237?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522755#comment-17522755
 ] 

dalongliu commented on FLINK-27237:
---

We have discussed and voted this ticket in 
[FLIP-63|[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support].]
 So it is the continuation of FLIP-63.

> Partitioned table statement enhancement
> ---
>
> Key: FLINK-27237
> URL: https://issues.apache.org/jira/browse/FLINK-27237
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: dalongliu
>Priority: Major
> Fix For: 1.16.0
>
>
> This is an umbrella issue which is used to discuss the syntax enhancement 
> about partitioned table. These new syntaxes are very useful for partitioned 
> tables, especially for batch job.
> Therefore, I propose to support the following statement about partitioned 
> table:
> {code:sql}
> -- add partition
> ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
> [PARTITION  ...] } 
> -- drop partition
> ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
> [PARTITION  ...] } [PURGE]
> -- rename partition
> ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
> ;
> -- show partitions
> SHOW PARTITIONS table_name [PARTITION ]
> :: 
>   (partition_col_name=partition_col_val, ...)
>  {code}
>  
> Reference:
> [1]: 
> [https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]
> [2]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]
> [3]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[jira] [Comment Edited] (FLINK-27237) Partitioned table statement enhancement

2022-04-15 Thread dalongliu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27237?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522755#comment-17522755
 ] 

dalongliu edited comment on FLINK-27237 at 4/15/22 9:29 AM:


We have discussed and voted this ticket in 
[FLIP-63|[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support]]
 So it is the continuation of FLIP-63.


was (Author: lsy):
We have discussed and voted this ticket in 
[FLIP-63|[https://cwiki.apache.org/confluence/display/FLINK/FLIP-63%3A+Rework+table+partition+support].]
 So it is the continuation of FLIP-63.

> Partitioned table statement enhancement
> ---
>
> Key: FLINK-27237
> URL: https://issues.apache.org/jira/browse/FLINK-27237
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: dalongliu
>Priority: Major
> Fix For: 1.16.0
>
>
> This is an umbrella issue which is used to discuss the syntax enhancement 
> about partitioned table. These new syntaxes are very useful for partitioned 
> tables, especially for batch job.
> Therefore, I propose to support the following statement about partitioned 
> table:
> {code:sql}
> -- add partition
> ALTER TABLE table_name ADD [IF NOT EXISTS] { PARTITION  
> [PARTITION  ...] } 
> -- drop partition
> ALTER TABLE table_name DROP [ IF EXISTS ] { PARTITION  
> [PARTITION  ...] } [PURGE]
> -- rename partition
> ALTER TABLE table_name PARTITION  RENAME TO PARTITION 
> ;
> -- show partitions
> SHOW PARTITIONS table_name [PARTITION ]
> :: 
>   (partition_col_name=partition_col_val, ...)
>  {code}
>  
> Reference:
> [1]: 
> [https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterPartition]
> [2]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-ddl-alter-table.html#add-partition|https://spark.apache.org/docs/3.0.0/sql-ref-syntax-ddl-alter-table.html#add-partition]
> [3]: 
> [https://spark.apache.org/docs/3.2.1/sql-ref-syntax-aux-show-partitions.html]



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink-kubernetes-operator] gyfora commented on pull request #165: [FLINK-26140] Support rollback strategies

2022-04-15 Thread GitBox


gyfora commented on PR #165:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/165#issuecomment-102022

   @wangyang0918 thanks, I will work on fixing this and covering with tests. 
Also I have a few follow up Jiras in mind including the sessionjob support 
   
   I am traveling today so I might need a day or so 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] wuchong merged pull request #19483: [hotfix][docs-zh] Fix "Google Cloud PubSub" Chinese page under "DataStream Connectors"

2022-04-15 Thread GitBox


wuchong merged PR #19483:
URL: https://github.com/apache/flink/pull/19483


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-27261) Disable web.cancel.enable for session jobs

2022-04-15 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-27261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17522753#comment-17522753
 ] 

Gyula Fora commented on FLINK-27261:


I don’t think we can do this on a per job level and session clusters can have 
non-managed jobs as well 

> Disable web.cancel.enable for session jobs
> --
>
> Key: FLINK-27261
> URL: https://issues.apache.org/jira/browse/FLINK-27261
> Project: Flink
>  Issue Type: Sub-task
>  Components: Kubernetes Operator
>Reporter: Yang Wang
>Priority: Major
>  Labels: starter
>
> In FLINK-27154, we disable {{web.cancel.enable}} for application cluster. We 
> should also do this for session jobs.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)


[GitHub] [flink] snailHumming commented on pull request #19488: [FLINK-25716][docs-zh] Apply CR changes

2022-04-15 Thread GitBox


snailHumming commented on PR #19488:
URL: https://github.com/apache/flink/pull/19488#issuecomment-1099983403

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



  1   2   >