[jira] [Comment Edited] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-12-13 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17248712#comment-17248712
 ] 

badqiu edited comment on FLINK-19146 at 12/14/20, 2:00 AM:
---

set `table.exec.mini-batch.enabled`=`true`;
 set `table.exec.mini-batch.allow-latency`=`5s`;
 set `table.exec.mini-batch.size`=`2000`;

 

CREATE TABLE test_table (
 name STRING
 ) WITH (
 'connector'='datagen',
 'rows-per-second'='1',
 'fields.name.length'='1'
 );

 

select name,count(1)
 from test_table
 group by name;

 

this select sql has no output。
 modify 

set `table.exec.mini-batch.size`=`2`;

will output result;


was (Author: badqiu):
set `table.exec.mini-batch.enabled`=`true`;
set `table.exec.mini-batch.allow-latency`=`5s`;
set `table.exec.mini-batch.size`=`2000`;

 

CREATE TABLE test_table (
name STRING
) WITH (
'connector'='datagen',
'rows-per-second'='1',
'fields.name.length'='1'
);

 

select name,count(*)
from test_table
group by name;

 

this select sql has no output。
modify 

set `table.exec.mini-batch.size`=`2`;

will output result;

> createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and 
> table.exec.mini-batch.allow-latency 
> 
>
> Key: FLINK-19146
> URL: https://issues.apache.org/jira/browse/FLINK-19146
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
> Attachments: mini_batch_trigger_by_latency.png, 
> mini_batch_trigger_by_size.png
>
>
> Using *or* conditions, you can control the total data delay and improve 
> computing performance.
> Increase the batch size to very large, but the data delay is still within the 
> set range.
>  
>  
> table.exec.mini-batch.size is true
> =>
> (table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-12-13 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17248712#comment-17248712
 ] 

badqiu commented on FLINK-19146:


set `table.exec.mini-batch.enabled`=`true`;
set `table.exec.mini-batch.allow-latency`=`5s`;
set `table.exec.mini-batch.size`=`2000`;

 

CREATE TABLE test_table (
name STRING
) WITH (
'connector'='datagen',
'rows-per-second'='1',
'fields.name.length'='1'
);

 

select name,count(*)
from test_table
group by name;

 

this select sql has no output。
modify 

set `table.exec.mini-batch.size`=`2`;

will output result;

> createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and 
> table.exec.mini-batch.allow-latency 
> 
>
> Key: FLINK-19146
> URL: https://issues.apache.org/jira/browse/FLINK-19146
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
> Attachments: mini_batch_trigger_by_latency.png, 
> mini_batch_trigger_by_size.png
>
>
> Using *or* conditions, you can control the total data delay and improve 
> computing performance.
> Increase the batch size to very large, but the data delay is still within the 
> set range.
>  
>  
> table.exec.mini-batch.size is true
> =>
> (table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-12-10 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17247717#comment-17247717
 ] 

badqiu commented on FLINK-19146:


As you said, it seems to work. But the actual test found no effect. Only size 
is in effect.

If the output result of my "group by" is less than 100, mini batch size=100. It 
will not any output

> createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and 
> table.exec.mini-batch.allow-latency 
> 
>
> Key: FLINK-19146
> URL: https://issues.apache.org/jira/browse/FLINK-19146
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
> Attachments: mini_batch_trigger_by_latency.png, 
> mini_batch_trigger_by_size.png
>
>
> Using *or* conditions, you can control the total data delay and improve 
> computing performance.
> Increase the batch size to very large, but the data delay is still within the 
> set range.
>  
>  
> table.exec.mini-batch.size is true
> =>
> (table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-12-10 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17247607#comment-17247607
 ] 

badqiu commented on FLINK-19146:


This problem is repeated if the output result set is small.

> createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and 
> table.exec.mini-batch.allow-latency 
> 
>
> Key: FLINK-19146
> URL: https://issues.apache.org/jira/browse/FLINK-19146
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
> Attachments: mini_batch_trigger_by_latency.png, 
> mini_batch_trigger_by_size.png
>
>
> Using *or* conditions, you can control the total data delay and improve 
> computing performance.
> Increase the batch size to very large, but the data delay is still within the 
> set range.
>  
>  
> table.exec.mini-batch.size is true
> =>
> (table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19390) mysql jdbc join error java.lang.ClassCastException: java.math.BigInteger cannot be cast to java.lang.Long

2020-09-24 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17201448#comment-17201448
 ] 

badqiu commented on FLINK-19390:


yes, But why use BigInteger, convert it to avoid errors。

> mysql jdbc join error java.lang.ClassCastException: java.math.BigInteger 
> cannot be cast to java.lang.Long
> -
>
> Key: FLINK-19390
> URL: https://issues.apache.org/jira/browse/FLINK-19390
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC, Table SQL / Ecosystem
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
>
> join on by   t1.order_id = t2.order_id.
> and order_id dataType: *bigint*.  will error.
>  
>  
> Caused by: java.lang.RuntimeException: java.lang.ClassCastException: 
> java.math.BigInteger cannot be cast to java.lang.Long
> at 
> org.apache.flink.table.runtime.operators.join.lookup.LookupJoinWithCalcRunner$CalcCollector.collect(LookupJoinWithCalcRunner.java:84)
> at 
> org.apache.flink.table.runtime.operators.join.lookup.LookupJoinWithCalcRunner$CalcCollector.collect(LookupJoinWithCalcRunner.java:71)
> at 
> org.apache.flink.table.functions.TableFunction.collect(TableFunction.java:203)
> at 
> org.apache.flink.connector.jdbc.table.JdbcRowDataLookupFunction.eval(JdbcRowDataLookupFunction.java:162)
> at LookupFunction$43.flatMap(Unknown Source)
> at 
> org.apache.flink.table.runtime.operators.join.lookup.LookupJoinRunner.processElement(LookupJoinRunner.java:82)
> at 
> org.apache.flink.table.runtime.operators.join.lookup.LookupJoinRunner.processElement(LookupJoinRunner.java:36)
> at 
> org.apache.flink.streaming.api.operators.ProcessOperator.processElement(ProcessOperator.java:66)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at StreamExecCalc$40.processElement(Unknown Source)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collect(StreamSourceContexts.java:104)
> at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collectWithTimestamp(StreamSourceContexts.java:111)
> at 
> org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:352)
> at 
> org.apache.flink.streaming.connectors.kafka.internal.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:185)
> at 
> org.apache.flink.streaming.connectors.kafka.internal.KafkaFetcher.runFetchLoop(KafkaFetcher.java:141)
> at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:755)
> at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:100)
> at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:63)
> at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:201)
> Caused by: java.lang.ClassCastException: java.math.BigInteger cannot be cast 
> to java.lang.Long
> at 
> org.apache.flink.table.data.GenericRowData.getLong(GenericRowData.java:154)
> at TableCalcMapFunction$47.flatMap(Unknown Source)
> at 
> org.apache.flink.table.runtime.operators.join.lookup.LookupJoinWithCalcRunner$CalcCollector.collect(LookupJoinWithCalcRunner.java:82)
> ... 27 more



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19390) mysql jdbc join error java.lang.ClassCastException: java.math.BigInteger cannot be cast to java.lang.Long

2020-09-24 Thread badqiu (Jira)
badqiu created FLINK-19390:
--

 Summary: mysql jdbc join error java.lang.ClassCastException: 
java.math.BigInteger cannot be cast to java.lang.Long
 Key: FLINK-19390
 URL: https://issues.apache.org/jira/browse/FLINK-19390
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.11.1
Reporter: badqiu


join on by   t1.order_id = t2.order_id.

and order_id dataType: *bigint*.  will error.

 

 
Caused by: java.lang.RuntimeException: java.lang.ClassCastException: 
java.math.BigInteger cannot be cast to java.lang.Long
at 
org.apache.flink.table.runtime.operators.join.lookup.LookupJoinWithCalcRunner$CalcCollector.collect(LookupJoinWithCalcRunner.java:84)
at 
org.apache.flink.table.runtime.operators.join.lookup.LookupJoinWithCalcRunner$CalcCollector.collect(LookupJoinWithCalcRunner.java:71)
at 
org.apache.flink.table.functions.TableFunction.collect(TableFunction.java:203)
at 
org.apache.flink.connector.jdbc.table.JdbcRowDataLookupFunction.eval(JdbcRowDataLookupFunction.java:162)
at LookupFunction$43.flatMap(Unknown Source)
at 
org.apache.flink.table.runtime.operators.join.lookup.LookupJoinRunner.processElement(LookupJoinRunner.java:82)
at 
org.apache.flink.table.runtime.operators.join.lookup.LookupJoinRunner.processElement(LookupJoinRunner.java:36)
at 
org.apache.flink.streaming.api.operators.ProcessOperator.processElement(ProcessOperator.java:66)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
at StreamExecCalc$40.processElement(Unknown Source)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
at 
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
at 
org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collect(StreamSourceContexts.java:104)
at 
org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collectWithTimestamp(StreamSourceContexts.java:111)
at 
org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:352)
at 
org.apache.flink.streaming.connectors.kafka.internal.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:185)
at 
org.apache.flink.streaming.connectors.kafka.internal.KafkaFetcher.runFetchLoop(KafkaFetcher.java:141)
at 
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:755)
at 
org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:100)
at 
org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:63)
at 
org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:201)
Caused by: java.lang.ClassCastException: java.math.BigInteger cannot be cast to 
java.lang.Long
at 
org.apache.flink.table.data.GenericRowData.getLong(GenericRowData.java:154)
at TableCalcMapFunction$47.flatMap(Unknown Source)
at 
org.apache.flink.table.runtime.operators.join.lookup.LookupJoinWithCalcRunner$CalcCollector.collect(LookupJoinWithCalcRunner.java:82)
... 27 more



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-09-07 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17191534#comment-17191534
 ] 

badqiu commented on FLINK-19146:


My test result is not like this.

 

search result : *trigger by size* 

!mini_batch_trigger_by_size.png!

 

 

search result : *trigger by latency*

!mini_batch_trigger_by_latency.png!

 

 

 

> createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and 
> table.exec.mini-batch.allow-latency 
> 
>
> Key: FLINK-19146
> URL: https://issues.apache.org/jira/browse/FLINK-19146
> Project: Flink
>  Issue Type: New Feature
>  Components: API / DataStream
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
> Attachments: mini_batch_trigger_by_latency.png, 
> mini_batch_trigger_by_size.png
>
>
> Using *or* conditions, you can control the total data delay and improve 
> computing performance.
> Increase the batch size to very large, but the data delay is still within the 
> set range.
>  
>  
> table.exec.mini-batch.size is true
> =>
> (table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-09-07 Thread badqiu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19146?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

badqiu updated FLINK-19146:
---
Attachment: mini_batch_trigger_by_size.png
mini_batch_trigger_by_latency.png

> createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and 
> table.exec.mini-batch.allow-latency 
> 
>
> Key: FLINK-19146
> URL: https://issues.apache.org/jira/browse/FLINK-19146
> Project: Flink
>  Issue Type: New Feature
>  Components: API / DataStream
>Affects Versions: 1.11.1
>Reporter: badqiu
>Priority: Major
> Attachments: mini_batch_trigger_by_latency.png, 
> mini_batch_trigger_by_size.png
>
>
> Using *or* conditions, you can control the total data delay and improve 
> computing performance.
> Increase the batch size to very large, but the data delay is still within the 
> set range.
>  
>  
> table.exec.mini-batch.size is true
> =>
> (table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19146) createMiniBatchTrigger() use OR ,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency

2020-09-07 Thread badqiu (Jira)
badqiu created FLINK-19146:
--

 Summary: createMiniBatchTrigger() use OR 
,table.exec.mini-batch.size and table.exec.mini-batch.allow-latency 
 Key: FLINK-19146
 URL: https://issues.apache.org/jira/browse/FLINK-19146
 Project: Flink
  Issue Type: New Feature
  Components: API / DataStream
Affects Versions: 1.11.1
Reporter: badqiu


Using *or* conditions, you can control the total data delay and improve 
computing performance.
Increase the batch size to very large, but the data delay is still within the 
set range.

 

 

table.exec.mini-batch.size is true

=>

(table.exec.mini-batch.size or table.exec.mini-batch.allow-latency) is true

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19048) support line format for table connector

2020-08-26 Thread badqiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19048?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17185081#comment-17185081
 ] 

badqiu commented on FLINK-19048:


 这个啥时候支持呢? 用这个 + UDF解析数据方便很多啊。

> support line format for table connector
> ---
>
> Key: FLINK-19048
> URL: https://issues.apache.org/jira/browse/FLINK-19048
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Ecosystem
>Reporter: badqiu
>Priority: Major
> Attachments: LineFormatFactory.java, 
> LineRowDataDeserializationSchema.java, LineRowDeserializationSchema.java, 
> LineRowFormatFactory.java, LineRowSchemaConverter.java, 
> LineRowSerializationSchema.java
>
>
> Native string data format. No data conversion is done.
> This format is particularly friendly to data without time attributes. With 
> UDF, writing your own data analysis will be much more convenient.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19048) support line format for table connector

2020-08-25 Thread badqiu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19048?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

badqiu updated FLINK-19048:
---
Summary: support line format for table connector  (was: suuport line format 
for table connector)

> support line format for table connector
> ---
>
> Key: FLINK-19048
> URL: https://issues.apache.org/jira/browse/FLINK-19048
> Project: Flink
>  Issue Type: New Feature
>  Components: API / Type Serialization System
>Reporter: badqiu
>Priority: Major
> Attachments: LineFormatFactory.java, 
> LineRowDataDeserializationSchema.java, LineRowDeserializationSchema.java, 
> LineRowFormatFactory.java, LineRowSchemaConverter.java, 
> LineRowSerializationSchema.java
>
>
> Native string data format. No data conversion is done.
> This format is particularly friendly to data without time attributes. With 
> UDF, writing your own data analysis will be much more convenient.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19048) suuport line format for table connector

2020-08-25 Thread badqiu (Jira)
badqiu created FLINK-19048:
--

 Summary: suuport line format for table connector
 Key: FLINK-19048
 URL: https://issues.apache.org/jira/browse/FLINK-19048
 Project: Flink
  Issue Type: New Feature
  Components: API / Type Serialization System
Reporter: badqiu
 Attachments: LineFormatFactory.java, 
LineRowDataDeserializationSchema.java, LineRowDeserializationSchema.java, 
LineRowFormatFactory.java, LineRowSchemaConverter.java, 
LineRowSerializationSchema.java

Native string data format. No data conversion is done.
This format is particularly friendly to data without time attributes. With UDF, 
writing your own data analysis will be much more convenient.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)