Re: Kafka to Flink to Hive - Writes failing

2018-06-11 Thread Piotr Nowojski
Yes, BucketingSink is a better option. You can start from looking at the BucketingSink java docs. Please also take a look on this: https://stackoverflow.com/questions/47669729/how-to-write-to-orc-files-using-bucketingsink-in-apache-flink

Re: Kafka to Flink to Hive - Writes failing

2018-06-10 Thread sagar loke
Thanks, We are getting data in Avro format from Kafka and are planning to write data in ORC format to Hive tables. 1. Is BucketingSink better option for this use case or something else ? 2. Is there a sample code example which we can refer ? Thanks in advance, On Sun, Jun 10, 2018 at 10:49 PM, J

Re: Kafka to Flink to Hive - Writes failing

2018-06-10 Thread Jörn Franke
Don’t use the JDBC driver to write to Hive. The performance of JDBC in general for large volumes is suboptimal. Write it to a file in HDFS in a format supported by HIve and point the table definition in Hive to it. > On 11. Jun 2018, at 04:47, sagar loke wrote: > > I am trying to Sink data to

Kafka to Flink to Hive - Writes failing

2018-06-10 Thread sagar loke
I am trying to Sink data to Hive via Confluent Kafka -> Flink -> Hive using following code snippet: But I am getting following error: final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); DataStream stream = readFromKafka(env); private static final TypeI