I also thought about it and my conclusion was to use a generic sql parser
(e.g. Calcite?) to extract the column names from the input query (because
in the query you can rename/add fields...).. I'd like to hear opinions
about this..unfortunately I don't have the time to implement this right now
:(

On Wed, Feb 8, 2017 at 1:59 PM, Punit Tandel <punit.tan...@ericsson.com>
wrote:

> HI
>
> With this approach i will be able to get data types but not the column
> names because TypeInformation<?> typeInformation = dataStream.getType()
> will return types but not the columns names.
>
> Is there any other way to get the column names from Row?
>
> Thanks
> Punit
>
> On 02/08/2017 10:17 AM, Chesnay Schepler wrote:
>
> Hello,
>
> in the JDBC case i would suggest that you extract the schema from the
> first Row that your sink receives, create the table, and then
> start writing data.
>
> However, keep in mind that Rows can contain null fields; so you may not be
> able to extract the entire schema if the first
> row has a null somewhere.
>
> Regards,
> Chesnay
>
> On 08.02.2017 10:48, Punit Tandel wrote:
>
> Hi Chesnay
>
> Currently that is what i have done, reading the schema from database in
> order to create a new table in jdbc database and writing the rows coming
> from jdbcinputformat.
>
> Overall i am trying to implement the solution which reads the streaming
> data from one source which either could be coming from kafka, Jdbc, Hive,
> Hdfs and writing those streaming data to output source which is again could
> be any of those.
>
> For a simple use case i have just taken one scenario using jdbc in and
> jdbc out, Since the jdbc input source returns the datastream of Row and to
> write them into jdbc database we have to create a table which requires
> schema.
>
> Thanks
> Punit
>
>
>
> On 02/08/2017 08:22 AM, Chesnay Schepler wrote:
>
> Hello,
>
> I don't understand why you explicitly need the schema since the batch
> JDBCInput-/Outputformats don't require it.
> That's kind of the nice thing about Rows.
>
> Would be cool if you could tell us what you're planning to do with the
> schema :)
>
> In any case, to get the schema within the plan then you will have to query
> the DB and build it yourself. Note that this
> is executed on the client.
>
> Regards,
> Chesnay
>
> On 08.02.2017 00:39, Punit Tandel wrote:
>
> Hi Robert
>
> Thanks for the response, So in near future release of the flink version ,
> is this functionality going to be implemented ?
>
> Thanks
> On 02/07/2017 04:12 PM, Robert Metzger wrote:
>
> Currently, there is no streaming JDBC connector.
> Check out this thread from last year: http://apache-flink-
> mailing-list-archive.1008284.n3.nabble.com/JDBC-Streaming-
> Connector-td10508.html
>
>
>
> On Mon, Feb 6, 2017 at 5:00 PM, Ufuk Celebi <u...@apache.org> wrote:
>
>> I'm not sure how well this works for the streaming API. Looping in
>> Chesnay, who worked on this.
>>
>> On Mon, Feb 6, 2017 at 11:09 AM, Punit Tandel <punit.tan...@ericsson.com>
>> wrote:
>> > Hi ,
>> >
>> > I was looking into flink streaming api and trying to implement the
>> solution
>> > for reading the data from jdbc database and writing them to jdbc databse
>> > again.
>> >
>> > At the moment i can see the datastream is returning Row from the
>> database.
>> > dataStream.getType().getGenericParameters() retuning an empty list of
>> > collection.
>> >
>> > I am right now manually creating a database connection and getting the
>> > schema from ResultMetadata and constructing the schema for the table
>> which
>> > is a bit heavy operation.
>> >
>> > So is there any other way to get the schema for the table in order to
>> create
>> > a new table and write those records in the database ?
>> >
>> > Please let me know
>> >
>> > Thanks
>> > Punit
>>
>
>
>
>
>
>
>

Reply via email to