hudi-bot opened a new issue, #16439:
URL: https://github.com/apache/hudi/issues/16439

   Currently the flink write task parallelism is set up through 
   {code:java}
   write.tasks{code}
   it is kind of a fixed number during the lifecycle of the ingestion pipeline, 
while for streaming, there are always fluctuation of the workload, it is great 
if we can tune the parallelism of write tasks based on the job load profile 
dynamically.
   
   On K8s, Flink provides a 
[https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/autoscaler/]
 which is suitable for the purpose, which deserves a further investigation.
   
   ## JIRA info
   
   - Link: https://issues.apache.org/jira/browse/HUDI-7574
   - Type: Improvement
   - Fix version(s):
     - 1.2.0
   
   
   ---
   
   
   ## Comments
   
   09/Apr/24 01:45;vinoth;We need to rethink these singleton tasks like 
cleaning etc. ;;;


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to