Ravi created FLINK-22830:
----------------------------
Summary: DataSet API to allow to collect and do the bulk upload in
DB
Key: FLINK-22830
URL: https://issues.apache.org/jira/browse/FLINK-22830
Project: Flink
Issue Type: Improvement
Reporter: Ravi
Experts,
I am trying to perform some ETL operation on large data set as batch
processing. My requirement is to extract the data , transform it and then save
to mongoDB. I am using Apache FLINK but the performance is very slow as I am
doing the mongoDB update on each row.
Is there any way where we can sink as bulk record so that performance can
increase. Like after all the transformation we do the bulk update on mongoDB.
We can aggregate them all and finally sink it in DB just like stream
[.aggregate() .sink(\{bulk update})] . Please refer
[https://stackoverflow.com/questions/67717964/java-apache-flink-batch-processing-performance-for-bulk-mongodb-update]
--
This message was sent by Atlassian Jira
(v8.3.4#803005)