veenaypatil opened a new issue #4017:
URL: https://github.com/apache/hudi/issues/4017


   **_Tips before filing an issue_**
   
   - Have you gone through our 
[FAQs](https://cwiki.apache.org/confluence/display/HUDI/FAQ)?
   
   - Join the mailing list to engage in conversations and get faster support at 
dev-subscr...@hudi.apache.org.
   
   - If you have triaged this as a bug, then file an 
[issue](https://issues.apache.org/jira/projects/HUDI/issues) directly.
   
   **Describe the problem you faced**
   
   We are getting the following error in Production for one of the end users 
ETL's
   ```
   Caused by: java.io.FileNotFoundException: No such file or directory: 
s3a://bucket/cdcv2/data/in_ums/user_umfnd_s3/2cf933ef-fe51-4e41-8b0d-af7fa5ed2d85-0_87-19419-8663185_20211116163235.parquet
   It is possible the underlying files have been updated. You can explicitly 
invalidate the cache in Spark by running 'REFRESH TABLE tableName' command in 
SQL or by recreating the Dataset/DataFrame involved.
   ```
   
   We had faced the same issue earlier but we mitigated it by increasing 
cleaner commits to 120 in spark streaming job which is writing to this 
location, For reference the spark streaming job has a batch interval of 10 mins 
where on an avg. the batches are completing in 4 mins and compaction takes 
40-50mins which is triggered after 4 commits, so roughly we have around 8hrs of 
commits. 
   
   User Is running the ETL on spark 2.x which is combination of Spark-SQL and  
Spark-core
   
   **To Reproduce**
   
   Steps to reproduce the behavior:
   
   1. We are consistently getting the same error even after retrying the ETL
   
   
   **Expected behavior**
   
   A clear and concise description of what you expected to happen.
   
   
   **Environment Description**
   
   * Hudi version : 0.8
   
   * Spark version : 2.x 
   
   * Hive version :  1.2.x
   
   * Hadoop version : 2.7
   
   * Storage (HDFS/S3/GCS..) : S3
   
   * Running on Docker? (yes/no) : no
   
   
   **Additional context**
   
   Add any other context about the problem here.
   
   The above configs are of older cluster where the ETL ran. 
   All other ETL's running on Spark3 and using Hive3 are running fine , as 
mentioned earlier where we had increased the cleaner commits, one of the ETL's 
had failed on newer cluster as well but post increasing the cleaner commits 
configs it has not failed on new cluster.
   
   **Stacktrace**
   
   ```
   Caused by: java.io.FileNotFoundException: No such file or directory: 
s3a://bucket/cdcv2/data/in_ums/user_umfnd_s3/2cf933ef-fe51-4e41-8b0d-af7fa5ed2d85-0_87-19419-8663185_20211116163235.parquet
   It is possible the underlying files have been updated. You can explicitly 
invalidate the cache in Spark by running 'REFRESH TABLE tableName' command in 
SQL or by recreating the Dataset/DataFrame involved.
   ```
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Reply via email to