Don’t use the JDBC driver to write to Hive. The performance of JDBC in general
for large volumes is suboptimal.
Write it to a file in HDFS in a format supported by HIve and point the table
definition in Hive to it.
> On 11. Jun 2018, at 04:47, sagar loke wrote:
>
> I am trying to Sink data to
after checking code, I found that BlobServer use ServerSocket instead of
Netty's ServerBootstrap.
I wonder why and is it ok to migtate to ServerBootstrap
--
Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/
I am trying to Sink data to Hive via Confluent Kafka -> Flink -> Hive using
following code snippet:
But I am getting following error:
final StreamExecutionEnvironment env =
StreamExecutionEnvironment.getExecutionEnvironment();
DataStream stream = readFromKafka(env);
private static final
Thank you very much Fabian I found that solution in the link below and this
is the bes fit for my use case
https://stackoverflow.com/questions/47458497/apache-flink-how-to-apply-multiple-counting-window-functions?rq=1
I am still testing how to Count (example. numTransactions >= 3) then I
sum I
I have a process function defined with these interfaces
public class MyProcessFunction extends ProcessFunction
implements CheckpointedFunction, ProcessingTimeCallback {...}
In snapshotState() method, I want to close files and emit the metadata
about the closed files to downstream operator.
Hi,
I'm trying to understand the state functionality. Is there any case where
using CheckpointedFunction with Keyed state makes any sense?
Or the CheckpointedFunction is only to be used with operator state?
Also, trying to understand the BufferinSink example here:
Hi Antonio,
Cascading window aggregations as done in your example is a good idea and is
preferable if the aggregation function is combinable, which is true for sum
(count can be done as sum of 1s).
Best, Fabian
2018-06-09 4:00 GMT+02:00 antonio saldivar :
> Hello
>
> Has anyone work this way?