Ravi created FLINK-22830:
----------------------------

             Summary: DataSet API to allow to collect and do the bulk upload in 
DB
                 Key: FLINK-22830
                 URL: https://issues.apache.org/jira/browse/FLINK-22830
             Project: Flink
          Issue Type: Improvement
            Reporter: Ravi


Experts,

 

I am trying to perform some ETL operation on large data set as batch 
processing. My requirement is to extract the data , transform it and then save 
to mongoDB. I am using Apache FLINK but the performance is very slow as I am 
doing the mongoDB update on each row.

Is there any way where we can sink as bulk record so that performance can 
increase. Like after all the transformation we do the bulk update on mongoDB. 
We can aggregate them all and finally sink it in DB just like stream 
[.aggregate() .sink(\{bulk update})]  . Please refer 
[https://stackoverflow.com/questions/67717964/java-apache-flink-batch-processing-performance-for-bulk-mongodb-update]
 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to