??null
-- --
??:
"user-zh"
2022-09-09 11:36:42,866 INFO
org.apache.flink.runtime.executiongraph.ExecutionGraph
[] - Source: HiveSource-ods_jt_hrs.ods_hrmis_HR_EMPL_Education (1/1)
(2a68412dab3602a1eeda5a750b308e23) switched from RUNNING to FAILED on
container_1658144991761_106260_01_02 @ hhny-cdh05 (dataPort=4
??Flink 1.13.2
HiveCatalogHive
Caused by: java.lang.RuntimeException: Hash join exceeded maximum number of
recursions, without reducing partitions enough to be memory resident. Probably
cause: Too many duplicate keys.
at
org
2022-01-23 04:31:39,568 INFO
org.apache.flink.runtime.executiongraph.ExecutionGraph
[] - Source: HiveSource-cosldatacenter.ods_rimdrill_dailyincidentsevents
-> Calc(select=[jobid, reportno, dayssincelast], where=[(idno = 1:BIGINT)])
(1/1) (7533d77baa7eb16e8242ae63e0706dff) switched from
>
> On Fri, Nov 19, 2021 at 10:48 AM Asahi Lee
<978466...@qq.com.invalid>
> wrote:
>
> > ??source
>
my_env/bin/activate??PYFLINK_CLIENT_EXECUTABLE??
> > jobmanagerNo module
rg/flink/flink-docs-release-1.14/docs/dev/python/python_config/#python-client-executable
??1.14.0??-pyclientexec
venv.zip/venv/bin/python
On Fri, Nov 19, 2021 at 10:48 AM Asahi Lee <978466...@qq.com.invalid> wrote:
"user-zh"
https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/dev/python/dependency_management/#python-interpreter-of-client
On Thu, Nov 18, 2021 at 9:00 PM Asahi Lee <978466...@qq.com.invalid> wrote:
> Hi !
"user-zh"
https://nightlies.apache.org/flink/flink-docs-release-1.12/downloads/setup-pyflink-virtual-env.sh
?? 2021-11-18 15:05:03??"Asahi Lee" <978466...@qq.com.INVALID> ??
>Hi!
>
flink??&n
Hi!
flink?? setup-pyflink-virtual-env.sh
python??
https://nightlies.apache.org/flink/flink-docs-release-1.14/zh/docs/dev/python/faq/
HI!
??flink 1.13.2??java table
apipython
udf??yarn-applicationyarn??pyflink?
hi!
??flink
sqlmysql??hive??yarn-application??16G??
hi! ??flink 1.13.2?? int
??avg?? int
double??decimal??bug
2021-09-06 11:20:32,787 WARN
org.apache.hadoop.hive.metastore.MetaStoreDirectSql: [pool-4-thread-192]:
Failed to execute [select "COLUMN_NAME", "COLUMN_TYPE", "LONG_LOW_VALUE",
"LONG_HIGH_VALUE", "DOUBLE_LOW_VALUE", "DOUBLE_HIGH_VALUE",
"BIG_DECIMAL_LOW_VALUE", "BI
Caused by: org.apache.flink.client.program.ProgramInvocationException: The main
method caused an error: SQL validation failed. Failed to get table column stats
of table cosldatacenter.ods_zyxx_coslzj_towing
at
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram
Caused by: java.lang.ArrayIndexOutOfBoundsException: 1024
at
org.apache.orc.impl.TreeReaderFactory$TreeReader.nextVector(TreeReaderFactory.java:269)
~[flink-sql-connector-hive-1.2.2_2.11-1.13.1.jar:1.13.1]
at
org.apache.orc.impl.TreeReaderFactory$TimestampTreeReader.nextVector(Tr
hi! ??flink 1.13.1??hivesql, ??
col1??string: ab'cd ,??instr '
??
instr(col1, '\'')
instr(col1, )
instr(col1, '\''')
instr(col1, '\\''')
hi!
??flink 1.13.1??hivedtsql hive??
Caused by:
org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException:
PartitionSpec CatalogPartitionSpec{{dt=2021-08-31}} does not match partition
keys [dt, xtlx, sblx] of table test_flink.test_par
hi!
flink??truncate table??flink
hivetruncate table??
hi??
??sqlselect0??hive
INSERT OVERWRITE target_table SELECT * from source_table where id > 10;
java.util.concurrent.CompletionException:
org.apache.flink.client.deployment.application.ApplicationExecutionException:
Could not execute application.
at
java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:292)
~[?:1.8.0_141]
at
java.util.concurrent.Co
SqlParser.parseStmtlist()sqlSqlNode
toString()??unicode
-- --
??:
"user-zh"
hive 1.1.0
-- --
??:
"user-zh"
hihive??sql??like
??
org.apache.flink.table.planner.codegen.CodeGenException: Unsupported call:
like(VARCHAR(255), STRING NOT NULL)
org.apache.flink.table.planner.codegen.CodeGenException: Unsupported call:
like(STRING, STRING NOT NULL)
If
CREATE TABLE `cosldatacenter.ods_emp_maindata_iadc_paramvalue`(
`paramvalue_id` string COMMENT '',
`platform_id` string COMMENT '',
`equipment_id` string COMMENT '',
`param_id` string COMMENT '',
`param_value` string COMMENT '',
`remark` string COMMENT '',
`create_time` string C
sql??sql
-- --
??:
"user-zh"
Hi??
??sql??Unnamed??bug??
??
node103.taskmanager.container_1627469681067_0030_01_02.e621b91ec4a34ababeb6b0e2c4d6f22b.Source:
HiveSource-qc_test_t_student_score -> Calc(select=[id,
CAST(_UTF-16LE'Bob':VARC
hi!
??else??sqlInvalid table alias or
column reference 'u' ??sql'u'
CREATE CATALOG `tempo_df_hive_default_catalog` WITH(
'type' = 'hive',
'default-database' = 'default'
);
USE CATALOG tempo_df_hive_default_catalog;
CREATE TABLE IF
Hi??
??flink 1.13.1??hive sql??
CREATE CATALOG `tempo_df_hive_default_catalog` WITH(
'type' = 'hive',
'default-database' = 'default'
);
USE CATALOG tempo_df_hive_default_catalog;
CREATE TABLE IF NOT EXISTS `default`.`tempo_blackhole_table` (
f0 INT
);
use cosldatacenter;
??flink 1.12.0??
-- --
??:
"user-zh"
hi!
??flink
1.13.1??sqlhive
??hive 2.3.6??flink-sql-connector-hive-2.3.6
package com.meritdata.cloud.flink.test;
import org.apache.flink.table.api.EnvironmentSettings;
import org.apache.flink.ta
hi??
1. flink 1.13.1
??row(a,b)bug??
2.
rowrowrowname??name
??
package test;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import or
hi??
org.apache.flink.table.catalog.Column
Serializable??Serializable??
https://flink-packages.org/packages/streaming-flink-dynamodb-connector
-- --
??:
"user-zh"
hi!
flink jdbc ??
table-name??
CREATE TABLE MyUserTable ( id BIGINT, name STRING, age INT, status
BOOLEAN, PRIMARY KEY (id) NOT ENFORCED ) WITH ('connector' = 'jdbc',
'url' = 'jdbc:mysql://localhost:3306/mydatabase','table-name' = 'o
hi??
flink
cep??cep
??
??flink
??
?? ??flink
1.12??flinkDataStream
API,??RuntimeExecutionMode.BATCH??
package com.meritdata.cloud.tempo.d
??
??
StreamExecutionEnvironment bsEnv =
StreamExecutionEnvironment.getExecutionEnvironment();
StreamTableEnvironment bsTableEnv = StreamTableEnvironment.create(bsEnv);
DataStream
??
??flink-1.12.0,??flinkmain??URLClassLoader??http://a.jar??jarrest
api jar/run??uber-jar,
job??
?? flink
1.12.0??org.apach.flink.table.runtime.generated.CompileUtils.compile()DataSteamTable
program cannot be compiled. This is a bug. Please file an
issue.?? public abstract java.lang.Object
org.apache.flink.api.java.fu
??
??join
-- --
??:
"user-zh"
?? Interval
Join??flink
1.11.2joinjoin??
l_table.l_rt = r_table.r_pt l_table.l_rt BETWEEN
r_table.r_pt - INTERVAL '10' SECOND AND
??format=json??flinkjacksonjackson??
-- --
??:
"Asah
?? ??flink sql
1.11.2??proctime()UTC??+08env.java.optsjvm??+08
json
BatchTableEnvironmenttable to dataset; dataset to table
-- --
??:
"user-zh"
??BatchTableEnvironment
-- --
??:
"user-zh"
??
??flink
1.11.2??blink??batch
// ** // BLINK BATCH QUERY // ** import
org.apache.flink.table.api.EnvironmentSettings; import
org.apache.flink.table.api.TableEnvironment; EnvironmentSettings bbSettings
e-1.11/ops/config.html#taskmanager-memory-jvm-metaspace-size
[2] https://issues.apache.org/jira/browse/FLINK-16681
Best,
Hailong Wang
?? 2020-11-04 19:08:37??"Asahi Lee" <978466...@qq.com> ??
>??
> ??flink
sql,JDBC?
??
??flink
sql,JDBC??mysql??OutOfMemoryError:
Metaspace.mysqlconnection??class??
??
flink??web
uisource??detailRecords
Sent??0??
?? flink
SQL??rebalance-- --
??: "zilong xiao"
??
??flink
sqlkafkasql??kafka1??2kafka
stream apisql
api???
??
StreamTableEnvironment.from("")??datagen??table??
??
package org.apache.flink.playgrounds.spendreport;
import org.apache.flink.streaming.
??
??StreamTableEnvironment.from("")
??package kafka;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import org
??sqljob??
-- --
??:
"user-zh"
??
insert
into??job
??
EnvironmentSettings bbSettings =
EnvironmentSettings.newInstance().useBlinkPlanner().build();
TableEnvironment bsTableEnv = TableEnvironment.create(bbSettings);
String sourceDDL = "CREATE TABLE datagen ( "
hello!
flink??iedaCannot have
more than one execute() or executeAsyc() call in a single
environmentjob
manager??
org.apache.flink.client.
1. package kafka;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import org.apache.flink.table.api.EnvironmentSettings;
import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
public class DataGenTest {
publ
-- --
??:
"user-zh"
https://issues.apache.org/jira/browse/FLINK-18665
<https://issues.apache.org/jira/browse/FLINK-18665>;
> ?? 2020??7??23????00:07??Asahi Lee <978466...
1.
StreamExecutionEnvironment bsEnv =
StreamExecutionEnvironment.getExecutionEnvironment();
EnvironmentSettings bsSettings =
EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
StreamTableEnvironment bsTableEnv =
StreamTableEnvironment.create(bsEnv
60 matches
Mail list logo