提供一下你的作业的SQL?详细一点的上下文 > -----原始邮件----- > 发件人: "junjie.m...@goupwith.com" <junjie.m...@goupwith.com> > 发送时间: 2022-11-02 15:21:39 (星期三) > 收件人: user-zh <user-zh@flink.apache.org> > 抄送: > 主题: flink1.14.5报错提示 “无法编译表程序。这是一个bug。请提交一个问题。” > > flink1.14.5遇到一个很奇怪的报错,报错提示 “无法编译表程序。这是一个bug。请提交一个问题。” > 报错信息如下: > /* 1 */ > /* 2 */ public final class WatermarkGenerator$253 > /* 3 */ extends > org.apache.flink.table.runtime.generated.WatermarkGenerator { > /* 4 */ > /* 5 */ private transient > org.apache.flink.table.runtime.typeutils.MapDataSerializer typeSerializer$255; > /* 6 */ > /* 7 */ public WatermarkGenerator$253(Object[] references) throws > Exception { > /* 8 */ typeSerializer$255 = > (((org.apache.flink.table.runtime.typeutils.MapDataSerializer) > references[0])); > /* 9 */ } > /* 10 */ > /* 11 */ @Override > /* 12 */ public void open(org.apache.flink.configuration.Configuration > parameters) throws Exception { > /* 13 */ > /* 14 */ } > /* 15 */ > /* 16 */ @Override > /* 17 */ public Long > currentWatermark(org.apache.flink.table.data.RowData row) throws Exception { > /* 18 */ > /* 19 */ org.apache.flink.table.data.MapData field$254; > /* 20 */ boolean isNull$254; > /* 21 */ org.apache.flink.table.data.MapData field$256; > /* 22 */ > /* 23 */ isNull$254 = row.isNullAt(2); > /* 24 */ field$254 = null; > /* 25 */ if (!isNull$254) { > /* 26 */ field$254 = row.getMap(2); > /* 27 */ } > /* 28 */ field$256 = field$254; > /* 29 */ if (!isNull$254) { > /* 30 */ field$256 = (org.apache.flink.table.data.MapData) > (typeSerializer$255.copy(field$256)); > /* 31 */ } > /* 32 */ > /* 33 */ > /* 34 */ if (isNull$254) { > /* 35 */ return null; > /* 36 */ } else { > /* 37 */ return field$256.getMillisecond(); > /* 38 */ } > /* 39 */ } > /* 40 */ > /* 41 */ @Override > /* 42 */ public void close() throws Exception { > /* 43 */ > /* 44 */ } > /* 45 */ } > /* 46 */ > > 2022-11-02 15:11:11 [Limit(offset=[0], fetch=[10]) -> > TableToDataSteam(type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false) -> Flat Map -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_120, > type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3) *ROWTIME*, `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f3]) -> TableToDataSteam(type=ROW<`f1` STRING, `f2` INT, `f3` > TIMESTAMP(3), `f4` VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` > ROW<`a` STRING, `b` INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` > MULTISET<STRING>> NOT NULL, rowtime=false) -> autoflow -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_125, > type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3) *ROWTIME*, `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f3]) -> Calc(select=[f6, f7, f8, f1, f2, f3, f4, f5.name AS > f5_name, f5.age AS f5_age, f5.info.a AS f5_info_a, f5.info.b AS f5_info_b]) > -> TableToDataSteam(type=ROW<`f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, > `f8` MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` > VARCHAR(32), `f5_name` STRING, `f5_age` INT, `f5_info_a` STRING, `f5_info_b` > INT> NOT NULL, rowtime=false) -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_130, > type=ROW<`f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` > MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` VARCHAR(32), > `f5_name` STRING, `f5_age` INT, `f5_info_a` STRING, `f5_info_b` INT> NOT > NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f8]) -> TableToDataSteam(type=ROW<`f6` MAP<STRING, STRING>, `f7` > ARRAY<STRING>, `f8` MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` > TIMESTAMP(3), `f4` VARCHAR(32), `f5_name` STRING, `f5_age` INT, `f5_info_a` > STRING, `f5_info_b` INT> NOT NULL, rowtime=false) -> autoflow (1/1)#0] WARN > o.a.f.t.r.generated.GeneratedClass - Failed to compile split code, falling > back to original code > org.apache.flink.util.FlinkRuntimeException: > org.apache.flink.api.common.InvalidProgramException: Table program cannot be > compiled. This is a bug. Please file an issue. > at > org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:76) > at > org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:98) > at > org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:69) > at > org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperatorFactory.createStreamOperator(WatermarkAssignerOperatorFactory.java:54) > at > org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:81) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperator(OperatorChain.java:712) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:686) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:187) > at > org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.<init>(RegularOperatorChain.java:63) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:666) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:654) > at > org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) > at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) > at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) > at java.lang.Thread.run(Thread.java:748) > Caused by: > org.apache.flink.shaded.guava30.com.google.common.util.concurrent.UncheckedExecutionException: > org.apache.flink.api.common.InvalidProgramException: Table program cannot be > compiled. This is a bug. Please file an issue. > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2051) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache.get(LocalCache.java:3962) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4859) > at > org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:74) > ... 38 common frames omitted > Caused by: org.apache.flink.api.common.InvalidProgramException: Table program > cannot be compiled. This is a bug. Please file an issue. > at > org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:89) > at > org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:74) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4864) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3529) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2278) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2155) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2045) > ... 41 common frames omitted > Caused by: org.codehaus.commons.compiler.CompileException: Line 37, Column > 45: A method named "getMillisecond" is not declared in any enclosing class > nor any supertype, nor through a static import > at org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12211) > at org.codehaus.janino.UnitCompiler.findIMethod(UnitCompiler.java:9084) > at org.codehaus.janino.UnitCompiler.compileGet2(UnitCompiler.java:5062) > at org.codehaus.janino.UnitCompiler.access$9100(UnitCompiler.java:215) > at > org.codehaus.janino.UnitCompiler$16.visitMethodInvocation(UnitCompiler.java:4423) > at > org.codehaus.janino.UnitCompiler$16.visitMethodInvocation(UnitCompiler.java:4396) > at org.codehaus.janino.Java$MethodInvocation.accept(Java.java:5073) > at org.codehaus.janino.UnitCompiler.compileGet(UnitCompiler.java:4396) > at org.codehaus.janino.UnitCompiler.compileGetValue(UnitCompiler.java:5662) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2649) > at org.codehaus.janino.UnitCompiler.access$2800(UnitCompiler.java:215) > at > org.codehaus.janino.UnitCompiler$6.visitReturnStatement(UnitCompiler.java:1504) > at > org.codehaus.janino.UnitCompiler$6.visitReturnStatement(UnitCompiler.java:1487) > at org.codehaus.janino.Java$ReturnStatement.accept(Java.java:3563) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1487) > at org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1567) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:1553) > at org.codehaus.janino.UnitCompiler.access$1700(UnitCompiler.java:215) > at org.codehaus.janino.UnitCompiler$6.visitBlock(UnitCompiler.java:1493) > at org.codehaus.janino.UnitCompiler$6.visitBlock(UnitCompiler.java:1487) > at org.codehaus.janino.Java$Block.accept(Java.java:2779) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1487) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2468) > at org.codehaus.janino.UnitCompiler.access$1900(UnitCompiler.java:215) > at org.codehaus.janino.UnitCompiler$6.visitIfStatement(UnitCompiler.java:1495) > at org.codehaus.janino.UnitCompiler$6.visitIfStatement(UnitCompiler.java:1487) > at org.codehaus.janino.Java$IfStatement.accept(Java.java:2950) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1487) > at org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1567) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:3388) > at > org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1357) > at > org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1330) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:822) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:432) > at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:215) > at > org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:411) > at > org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:406) > at > org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1414) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:406) > at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:378) > at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:237) > at > org.codehaus.janino.SimpleCompiler.compileToClassLoader(SimpleCompiler.java:465) > at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:216) > at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:207) > at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:80) > at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:75) > at > org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:86) > ... 47 common frames omitted > /* 1 */ > /* 2 */ public final class WatermarkGenerator$253 > /* 3 */ extends > org.apache.flink.table.runtime.generated.WatermarkGenerator { > /* 4 */ > /* 5 */ private transient > org.apache.flink.table.runtime.typeutils.MapDataSerializer typeSerializer$255; > /* 6 */ > /* 7 */ public WatermarkGenerator$253(Object[] references) throws > Exception { > /* 8 */ typeSerializer$255 = > (((org.apache.flink.table.runtime.typeutils.MapDataSerializer) > references[0])); > /* 9 */ } > /* 10 */ > /* 11 */ @Override > /* 12 */ public void open(org.apache.flink.configuration.Configuration > parameters) throws Exception { > /* 13 */ > /* 14 */ } > /* 15 */ > /* 16 */ @Override > /* 17 */ public Long > currentWatermark(org.apache.flink.table.data.RowData row) throws Exception { > /* 18 */ > /* 19 */ org.apache.flink.table.data.MapData field$254; > /* 20 */ boolean isNull$254; > /* 21 */ org.apache.flink.table.data.MapData field$256; > /* 22 */ > /* 23 */ isNull$254 = row.isNullAt(2); > /* 24 */ field$254 = null; > /* 25 */ if (!isNull$254) { > /* 26 */ field$254 = row.getMap(2); > /* 27 */ } > /* 28 */ field$256 = field$254; > /* 29 */ if (!isNull$254) { > /* 30 */ field$256 = (org.apache.flink.table.data.MapData) > (typeSerializer$255.copy(field$256)); > /* 31 */ } > /* 32 */ > /* 33 */ > /* 34 */ if (isNull$254) { > /* 35 */ return null; > /* 36 */ } else { > /* 37 */ return field$256.getMillisecond(); > /* 38 */ } > /* 39 */ } > /* 40 */ > /* 41 */ @Override > /* 42 */ public void close() throws Exception { > /* 43 */ > /* 44 */ } > /* 45 */ } > /* 46 */ > > 2022-11-02 15:11:11 [Limit(offset=[0], fetch=[10]) -> > TableToDataSteam(type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false) -> Flat Map -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_120, > type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3) *ROWTIME*, `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f3]) -> TableToDataSteam(type=ROW<`f1` STRING, `f2` INT, `f3` > TIMESTAMP(3), `f4` VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` > ROW<`a` STRING, `b` INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` > MULTISET<STRING>> NOT NULL, rowtime=false) -> autoflow -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_125, > type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3) *ROWTIME*, `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f3]) -> Calc(select=[f6, f7, f8, f1, f2, f3, f4, f5.name AS > f5_name, f5.age AS f5_age, f5.info.a AS f5_info_a, f5.info.b AS f5_info_b]) > -> TableToDataSteam(type=ROW<`f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, > `f8` MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` > VARCHAR(32), `f5_name` STRING, `f5_age` INT, `f5_info_a` STRING, `f5_info_b` > INT> NOT NULL, rowtime=false) -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_130, > type=ROW<`f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` > MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` VARCHAR(32), > `f5_name` STRING, `f5_age` INT, `f5_info_a` STRING, `f5_info_b` INT> NOT > NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f8]) -> TableToDataSteam(type=ROW<`f6` MAP<STRING, STRING>, `f7` > ARRAY<STRING>, `f8` MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` > TIMESTAMP(3), `f4` VARCHAR(32), `f5_name` STRING, `f5_age` INT, `f5_info_a` > STRING, `f5_info_b` INT> NOT NULL, rowtime=false) -> autoflow (1/1)#0] WARN > o.a.flink.runtime.taskmanager.Task - Limit(offset=[0], fetch=[10]) -> > TableToDataSteam(type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false) -> Flat Map -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_120, > type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3) *ROWTIME*, `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f3]) -> TableToDataSteam(type=ROW<`f1` STRING, `f2` INT, `f3` > TIMESTAMP(3), `f4` VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` > ROW<`a` STRING, `b` INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` > MULTISET<STRING>> NOT NULL, rowtime=false) -> autoflow -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_125, > type=ROW<`f1` STRING, `f2` INT, `f3` TIMESTAMP(3) *ROWTIME*, `f4` > VARCHAR(32), `f5` ROW<`name` STRING, `age` INT, `info` ROW<`a` STRING, `b` > INT>>, `f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` MULTISET<STRING>> > NOT NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f3]) -> Calc(select=[f6, f7, f8, f1, f2, f3, f4, f5.name AS > f5_name, f5.age AS f5_age, f5.info.a AS f5_info_a, f5.info.b AS f5_info_b]) > -> TableToDataSteam(type=ROW<`f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, > `f8` MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` > VARCHAR(32), `f5_name` STRING, `f5_age` INT, `f5_info_a` STRING, `f5_info_b` > INT> NOT NULL, rowtime=false) -> > DataSteamToTable(stream=default_catalog.default_database.Unregistered_DataStream_Source_130, > type=ROW<`f6` MAP<STRING, STRING>, `f7` ARRAY<STRING>, `f8` > MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` TIMESTAMP(3), `f4` VARCHAR(32), > `f5_name` STRING, `f5_age` INT, `f5_info_a` STRING, `f5_info_b` INT> NOT > NULL, rowtime=false, watermark=false) -> WatermarkAssigner(rowtime=[f3], > watermark=[f8]) -> TableToDataSteam(type=ROW<`f6` MAP<STRING, STRING>, `f7` > ARRAY<STRING>, `f8` MULTISET<STRING>, `f1` STRING, `f2` INT, `f3` > TIMESTAMP(3), `f4` VARCHAR(32), `f5_name` STRING, `f5_age` INT, `f5_info_a` > STRING, `f5_info_b` INT> NOT NULL, rowtime=false) -> autoflow (1/1)#0 > (cbdd2f5f1c1b9f8bd596f1dc8e2da6f2) switched from INITIALIZING to FAILED with > failure cause: java.lang.RuntimeException: Could not instantiate generated > class 'WatermarkGenerator$253' > at > org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:75) > at > org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperatorFactory.createStreamOperator(WatermarkAssignerOperatorFactory.java:54) > at > org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:81) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperator(OperatorChain.java:712) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:686) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:676) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:626) > at > org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:187) > at > org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.<init>(RegularOperatorChain.java:63) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:666) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:654) > at > org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) > at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) > at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) > at java.lang.Thread.run(Thread.java:748) > Caused by: org.apache.flink.util.FlinkRuntimeException: > org.apache.flink.api.common.InvalidProgramException: Table program cannot be > compiled. This is a bug. Please file an issue. > at > org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:76) > at > org.apache.flink.table.runtime.generated.GeneratedClass.compile(GeneratedClass.java:102) > at > org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:69) > ... 36 more > Caused by: > org.apache.flink.shaded.guava30.com.google.common.util.concurrent.UncheckedExecutionException: > org.apache.flink.api.common.InvalidProgramException: Table program cannot be > compiled. This is a bug. Please file an issue. > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2051) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache.get(LocalCache.java:3962) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4859) > at > org.apache.flink.table.runtime.generated.CompileUtils.compile(CompileUtils.java:74) > ... 38 more > Caused by: org.apache.flink.api.common.InvalidProgramException: Table program > cannot be compiled. This is a bug. Please file an issue. > at > org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:89) > at > org.apache.flink.table.runtime.generated.CompileUtils.lambda$compile$1(CompileUtils.java:74) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4864) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3529) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2278) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2155) > at > org.apache.flink.shaded.guava30.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2045) > ... 41 more > Caused by: org.codehaus.commons.compiler.CompileException: Line 37, Column > 45: A method named "getMillisecond" is not declared in any enclosing class > nor any supertype, nor through a static import > at org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12211) > at org.codehaus.janino.UnitCompiler.findIMethod(UnitCompiler.java:9084) > at org.codehaus.janino.UnitCompiler.compileGet2(UnitCompiler.java:5062) > at org.codehaus.janino.UnitCompiler.access$9100(UnitCompiler.java:215) > at > org.codehaus.janino.UnitCompiler$16.visitMethodInvocation(UnitCompiler.java:4423) > at > org.codehaus.janino.UnitCompiler$16.visitMethodInvocation(UnitCompiler.java:4396) > at org.codehaus.janino.Java$MethodInvocation.accept(Java.java:5073) > at org.codehaus.janino.UnitCompiler.compileGet(UnitCompiler.java:4396) > at org.codehaus.janino.UnitCompiler.compileGetValue(UnitCompiler.java:5662) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2649) > at org.codehaus.janino.UnitCompiler.access$2800(UnitCompiler.java:215) > at > org.codehaus.janino.UnitCompiler$6.visitReturnStatement(UnitCompiler.java:1504) > at > org.codehaus.janino.UnitCompiler$6.visitReturnStatement(UnitCompiler.java:1487) > at org.codehaus.janino.Java$ReturnStatement.accept(Java.java:3563) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1487) > at org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1567) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:1553) > at org.codehaus.janino.UnitCompiler.access$1700(UnitCompiler.java:215) > at org.codehaus.janino.UnitCompiler$6.visitBlock(UnitCompiler.java:1493) > at org.codehaus.janino.UnitCompiler$6.visitBlock(UnitCompiler.java:1487) > at org.codehaus.janino.Java$Block.accept(Java.java:2779) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1487) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:2468) > at org.codehaus.janino.UnitCompiler.access$1900(UnitCompiler.java:215) > at org.codehaus.janino.UnitCompiler$6.visitIfStatement(UnitCompiler.java:1495) > at org.codehaus.janino.UnitCompiler$6.visitIfStatement(UnitCompiler.java:1487) > at org.codehaus.janino.Java$IfStatement.accept(Java.java:2950) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:1487) > at org.codehaus.janino.UnitCompiler.compileStatements(UnitCompiler.java:1567) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:3388) > at > org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1357) > at > org.codehaus.janino.UnitCompiler.compileDeclaredMethods(UnitCompiler.java:1330) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:822) > at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:432) > at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:215) > at > org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:411) > at > org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclaration(UnitCompiler.java:406) > at > org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1414) > at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:406) > at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:378) > at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:237) > at > org.codehaus.janino.SimpleCompiler.compileToClassLoader(SimpleCompiler.java:465) > at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:216) > at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:207) > at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:80) > at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:75) > at > org.apache.flink.table.runtime.generated.CompileUtils.doCompile(CompileUtils.java:86) > ... 47 more
------------------------------ Best, Ron