不应该呀,我看到仍然有 def registerDataStream[T](name: String, dataStream: DataStream[T], fields: Expression*): Unit
这个方法的,你能提供完整一点的上下文和报错吗? Best, tison. ddwcg <3149768...@qq.com> 于2019年8月26日周一 上午11:38写道: > 感谢您的回复,确实是这个原因。但是后面注册表的时候不能使用 Expression > 总是感觉 java api 和scala api有点混乱了 > > > 在 2019年8月26日,11:22,Zili Chen <wander4...@gmail.com> 写道: > > 试试把 > > import > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment > > 换成 > > import org.apache.flink.table.api.scala.StreamExecutionEnvironment > > 应该是意外 import 了不同包下的同名类的缘故 > > Best, > tison. > > > ddwcg <3149768...@qq.com> 于2019年8月26日周一 上午11:12写道: > > 大家好, > 我周末升级到了1.9.0,但是在初始化table env的时候编译不通过,请大家帮忙看看是什么原因, > > import org.apache.flink.streaming.api.CheckpointingMode > import > org.apache.flink.streaming.api.environment.CheckpointConfig.ExternalizedCheckpointCleanup > import > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment > import org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer > import org.apache.flink.streaming.util.serialization.SimpleStringSchema > import org.apache.flink.table.api.scala.StreamTableEnvironment > import org.apache.flink.table.planner.expressions.StddevPop > import org.apache.kafka.clients.consumer.ConsumerConfig > import org.apache.kafka.clients.producer.ProducerConfig > > object StreamingJob { > def main(args: Array[String]) { > val kafkaTopic = "source.kafka.topic" > val jobName ="test" > val parallelism =1 > val checkPointPath ="checkpoint/" > val kafkaBrokers ="" > > // set up the streaming execution environment > val env = StreamExecutionEnvironment.getExecutionEnvironment > env.setParallelism(parallelism) > env.enableCheckpointing(10000) > > > env.getCheckpointConfig.setCheckpointingMode(CheckpointingMode.EXACTLY_ONCE) > env.getCheckpointConfig.setMaxConcurrentCheckpoints(1) > > > env.getCheckpointConfig.enableExternalizedCheckpoints(ExternalizedCheckpointCleanup.RETAIN_ON_CANCELLATION) > //env.setStateBackend(new FsStateBackend(checkPointPath)) > > > val tableEnv = StreamTableEnvironment.create(env) > > > 提示有多个实现: > > 下面是pom文件: > > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-scala_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>compile</scope> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-streaming-scala_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>compile</scope> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > > <artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > <!-- > https://mvnrepository.com/artifact/org.apache.flink/flink-table-runtime-blink > --> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-table-runtime-blink_2.11</artifactId> > <version>1.9.0</version> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-connector-kafka_2.11</artifactId> > <version>1.9.0</version> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-table-common</artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > > > > > >