[ 
https://issues.apache.org/jira/browse/FLINK-9852?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16546317#comment-16546317
 ] 

ASF GitHub Bot commented on FLINK-9852:
---------------------------------------

Github user pnowojski commented on a diff in the pull request:

    https://github.com/apache/flink/pull/6343#discussion_r202964715
  
    --- Diff: 
flink-libraries/flink-table/src/main/scala/org/apache/flink/table/descriptors/SchematicDescriptor.scala
 ---
    @@ -19,14 +19,17 @@
     package org.apache.flink.table.descriptors
     
     /**
    -  * Common class for all descriptors describing a table sink.
    +  * A trait for descriptors that allow to define a format and schema.
       */
    -abstract class TableSinkDescriptor extends TableDescriptor {
    +trait SchematicDescriptor extends Descriptor {
    --- End diff --
    
    are you using it anywhere as interface? what does extracting it to separate 
interface give us? Maybe drop it?


> Expose descriptor-based sink creation in table environments
> -----------------------------------------------------------
>
>                 Key: FLINK-9852
>                 URL: https://issues.apache.org/jira/browse/FLINK-9852
>             Project: Flink
>          Issue Type: New Feature
>          Components: Table API & SQL
>            Reporter: Timo Walther
>            Assignee: Timo Walther
>            Priority: Major
>              Labels: pull-request-available
>
> Currently, only a table source can be created using the unified table 
> descriptors with {{tableEnv.from(...)}}. A similar approach should be 
> supported for defining sinks or even both types at the same time.
> I suggest the following syntax:
> {code}
> tableEnv.connect(Kafka(...)).registerSource("name")
> tableEnv.connect(Kafka(...)).registerSink("name")
> tableEnv.connect(Kafka(...)).registerSourceAndSink("name")
> {code}
> A table could then access the registered source/sink.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to