Hi Aakash, I believe there were schema level changes from Hudi 0.5.0 to 0.6.0 regarding those commit files. So if you are jumping from 0.5.0 to 0.8.0 right away, you will likely experience such an error, i.e. Failed to archive commits. You shouldn't need to delete archived files; instead, you should try deleting some, if not all, active commit files under your *.hoodie* folder. The reason for that is 0.8.0 is using a new AVRO schema to parse your old commit files, so you got the failure. Can you try the above approach and let us know? Thank you. :)
Best, Susu On Wed, Jun 23, 2021 at 12:21 PM aakash aakash <email2aak...@gmail.com> wrote: > Hi, > > I am trying to use Hudi 0.8 with Spark 3.0 in my prod environment and > earlier we were running Hudi 0.5 with Spark 2.4.4. > > While updating a very old index, I am getting this error : > > *from the logs it seem its error out while reading this file : > hudi/.hoodie/archived/.commits_.archive.119_1-0-1 in s3* > > 21/06/22 19:18:06 ERROR HoodieTimelineArchiveLog: Failed to archive > commits, .commit file: 20200715192915.rollback.inflight > java.io.IOException: Not an Avro data file > at org.apache.avro.file.DataFileReader.openReader(DataFileReader.java:50) > at > > org.apache.hudi.common.table.timeline.TimelineMetadataUtils.deserializeAvroMetadata(TimelineMetadataUtils.java:175) > at > > org.apache.hudi.client.utils.MetadataConversionUtils.createMetaWrapper(MetadataConversionUtils.java:84) > at > > org.apache.hudi.table.HoodieTimelineArchiveLog.convertToAvroRecord(HoodieTimelineArchiveLog.java:370) > at > > org.apache.hudi.table.HoodieTimelineArchiveLog.archive(HoodieTimelineArchiveLog.java:311) > at > > org.apache.hudi.table.HoodieTimelineArchiveLog.archiveIfRequired(HoodieTimelineArchiveLog.java:128) > at > > org.apache.hudi.client.AbstractHoodieWriteClient.postCommit(AbstractHoodieWriteClient.java:430) > at > > org.apache.hudi.client.AbstractHoodieWriteClient.commitStats(AbstractHoodieWriteClient.java:186) > at > > org.apache.hudi.client.SparkRDDWriteClient.commit(SparkRDDWriteClient.java:121) > at > > org.apache.hudi.HoodieSparkSqlWriter$.commitAndPerformPostOperations(HoodieSparkSqlWriter.scala:479) > > > Is this a backward compatibility issue? I have deleted a few archive files > but the problem is persisting so it does not look like a file corruption > issue. > > Regards, > Aakash >