[ 
https://issues.apache.org/jira/browse/FLINK-22830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arvid Heise closed FLINK-22830.
-------------------------------
    Resolution: Not A Problem

> DataSet API to allow to collect and do the bulk upload in DB
> ------------------------------------------------------------
>
>                 Key: FLINK-22830
>                 URL: https://issues.apache.org/jira/browse/FLINK-22830
>             Project: Flink
>          Issue Type: Improvement
>          Components: API / DataSet
>            Reporter: Ravi
>            Priority: Major
>
> Experts,
>  
> I am trying to perform some ETL operation on large data set as batch 
> processing. My requirement is to extract the data , transform it and then 
> save to mongoDB. I am using Apache FLINK but the performance is very slow as 
> I am doing the mongoDB update on each row.
> Is there any way where we can sink as bulk record so that performance can 
> increase. Like after all the transformation we do the bulk update on mongoDB. 
> We can aggregate them all and finally sink it in DB just like stream 
> [.aggregate() .sink(\{bulk update})]  . Please refer 
> [https://stackoverflow.com/questions/67717964/java-apache-flink-batch-processing-performance-for-bulk-mongodb-update]
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to