hudi-bot opened a new issue, #14714:
URL: https://github.com/apache/hudi/issues/14714

   https://travis-ci.com/github/apache/hudi/jobs/458936905
   
   [INFO] Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 
19.245 s - in org.apache.hudi.table.action.compact.TestInlineCompaction[INFO] 
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.245 s - in 
org.apache.hudi.table.action.compact.TestInlineCompaction[INFO] Running 
org.apache.hudi.table.action.compact.TestAsyncCompaction[WARN ] 2020-12-12 
15:13:43,814 org.apache.hudi.testutils.HoodieClientTestHarness  - Closing 
file-system instance used in previous test-run[WARN ] 2020-12-12 15:13:50,370 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[WARN ] 2020-12-12 15:14:02,285 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[WARN ] 2020-12-12 15:14:08,596 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[WARN ] 2020-12-12 15:14:16,857 
org.apache.hudi.common.util.ClusteringUtils  - 
 No content found in requested file for instant 
[==>006__replacecommit__REQUESTED][WARN ] 2020-12-12 15:14:16,861 
org.apache.hudi.common.util.ClusteringUtils  - No content found in requested 
file for instant [==>006__replacecommit__REQUESTED][ERROR] 2020-12-12 
15:14:16,919 org.apache.hudi.timeline.service.FileSystemViewHandler  - Got 
runtime exception servicing request 
partition=2015%2F03%2F17&basepath=%2Ftmp%2Fjunit7781027189613842524%2Fdataset&lastinstantts=005&timelinehash=ba1d2bb94a4b1d1e6e294e77086957b6c7c43b5a306e36cba6bbaa955a0ed8ceorg.apache.hudi.exception.HoodieIOException:
 Error reading clustering plan 006 at 
org.apache.hudi.common.util.ClusteringUtils.getClusteringPlan(ClusteringUtils.java:85)
 at 
org.apache.hudi.common.util.ClusteringUtils.lambda$getAllPendingClusteringPlans$0(ClusteringUtils.java:67)
 at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) 
at 
java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1374) 
at java.util.
 stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) at 
java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) at 
org.apache.hudi.common.util.ClusteringUtils.getAllFileGroupsInPendingClusteringPlans(ClusteringUtils.java:100)
 at 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.init(AbstractTableFileSystemView.java:111)
 at 
org.apache.hudi.common.table.view.RocksDbBasedFileSystemView.init(RocksDbBasedFileSystemView.java:91)
 at 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.runSync(AbstractTableFileSystemView.java:1077)
 at 
org.apache.hudi.common.table.view.IncrementalTimelineSyncFileSystemView.runSync(IncrementalTimelineSyncFileSystemView.java:97)
 at org.apache.hudi.common.table.view.AbstractTableFileSystemVie
 w.sync(AbstractTableFileSystemView.java:1059) at 
org.apache.hudi.timeline.service.FileSystemViewHandler.syncIfLocalViewBehind(FileSystemViewHandler.java:124)
 at 
org.apache.hudi.timeline.service.FileSystemViewHandler.access$100(FileSystemViewHandler.java:55)
 at 
org.apache.hudi.timeline.service.FileSystemViewHandler$ViewHandler.handle(FileSystemViewHandler.java:338)
 at io.javalin.security.SecurityUtil.noopAccessManager(SecurityUtil.kt:22) at 
io.javalin.Javalin.lambda$addHandler$0(Javalin.java:606) at 
io.javalin.core.JavalinServlet$service$2$1.invoke(JavalinServlet.kt:46) at 
io.javalin.core.JavalinServlet$service$2$1.invoke(JavalinServlet.kt:17) at 
io.javalin.core.JavalinServlet$service$1.invoke(JavalinServlet.kt:143) at 
io.javalin.core.JavalinServlet$service$2.invoke(JavalinServlet.kt:41) at 
io.javalin.core.JavalinServlet.service(JavalinServlet.kt:107) at 
io.javalin.core.util.JettyServerUtil$initialize$httpHandler$1.doHandle(JettyServerUtil.kt:72)
 at org.eclipse.jetty.server.handler.S
 copedHandler.nextScope(ScopedHandler.java:203) at 
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:480) at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1668)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:201)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1247)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:144) 
at org.eclipse.jetty.server.handler.HandlerList.handle(HandlerList.java:61) at 
org.eclipse.jetty.server.handler.StatisticsHandler.handle(StatisticsHandler.java:174)
 at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132) 
at org.eclipse.jetty.server.Server.handle(Server.java:502) at 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:370) at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:267) at 
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnectio
 n.java:305) at 
org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:103) at 
org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:117) at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:333)
 at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:310)
 at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:168)
 at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:126)
 at 
org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:366)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:765)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:683) 
at java.lang.Thread.run(Thread.java:748)Caused by: java.io.IOException: Not an 
Avro data file at 
org.apache.avro.file.DataFileReader.openReader(DataFileReader.java:63) at 
org.apache.hudi.common.table.timeline
 .TimelineMetadataUtils.deserializeAvroMetadata(TimelineMetadataUtils.java:160) 
at 
org.apache.hudi.common.table.timeline.TimelineMetadataUtils.deserializeRequestedReplaceMetadta(TimelineMetadataUtils.java:154)
 at 
org.apache.hudi.common.util.ClusteringUtils.getClusteringPlan(ClusteringUtils.java:79)
 ... 48 more[ERROR] 2020-12-12 15:14:16,933 
org.apache.hudi.common.table.view.PriorityBasedFileSystemView  - Got error 
running preferred function. Trying 
secondaryorg.apache.hudi.exception.HoodieRemoteException: Server Error at 
org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.getLatestFileSlices(RemoteHoodieTableFileSystemView.java:279)
 at 
org.apache.hudi.common.table.view.PriorityBasedFileSystemView.execute(PriorityBasedFileSystemView.java:81)
 at 
org.apache.hudi.common.table.view.PriorityBasedFileSystemView.getLatestFileSlices(PriorityBasedFileSystemView.java:160)
 at 
org.apache.hudi.table.action.commit.SparkInsertOverwriteCommitActionExecutor.getAllExistingFileIds(SparkIn
 sertOverwriteCommitActionExecutor.java:82) at 
org.apache.hudi.table.action.commit.SparkInsertOverwriteCommitActionExecutor.lambda$getPartitionToReplacedFileIds$81998308$1(SparkInsertOverwriteCommitActionExecutor.java:77)
 at 
org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1043)
 at 
org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1043)
 at scala.collection.Iterator$$anon$11.next(Iterator.scala:410) at 
scala.collection.Iterator$class.foreach(Iterator.scala:891) at 
scala.collection.AbstractIterator.foreach(Iterator.scala:1334) at 
scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59) at 
scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104) at 
scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48) at 
scala.collection.TraversableOnce$class.to(TraversableOnce.scala:310) at 
scala.collection.AbstractIterator.to(Iterator.scala:1334) at scala.collection.Tr
 aversableOnce$class.toBuffer(TraversableOnce.scala:302) at 
scala.collection.AbstractIterator.toBuffer(Iterator.scala:1334) at 
scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:289) at 
scala.collection.AbstractIterator.toArray(Iterator.scala:1334) at 
org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$13.apply(RDD.scala:945) at 
org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$13.apply(RDD.scala:945) at 
org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) 
at 
org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) 
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) at 
org.apache.spark.scheduler.Task.run(Task.scala:123) at 
org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:408)
 at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360) at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414) at 
java.util.concurrent.ThreadPoolExecutor.runWorker(
 ThreadPoolExecutor.java:1149) at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
at java.lang.Thread.run(Thread.java:748)Caused by: 
org.apache.http.client.HttpResponseException: Server Error at 
org.apache.http.impl.client.AbstractResponseHandler.handleResponse(AbstractResponseHandler.java:69)
 at org.apache.http.client.fluent.Response.handleResponse(Response.java:90) at 
org.apache.http.client.fluent.Response.returnContent(Response.java:97) at 
org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.executeRequest(RemoteHoodieTableFileSystemView.java:173)
 at 
org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.getLatestFileSlices(RemoteHoodieTableFileSystemView.java:275)
 ... 30 more[WARN ] 2020-12-12 15:14:18,248 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[WARN ] 2020-12-12 15:14:26,284 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance
  used in previous test-run[WARN ] 2020-12-12 15:14:37,355 
org.apache.hudi.table.action.rollback.BaseRollbackActionExecutor  - Rollback 
finished without deleting inflight instant file. 
Instant=[==>005__compaction__INFLIGHT][WARN ] 2020-12-12 15:14:38,244 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[WARN ] 2020-12-12 15:14:46,809 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[WARN ] 2020-12-12 15:14:53,337 
org.apache.hudi.table.action.rollback.BaseRollbackActionExecutor  - Rollback 
finished without deleting inflight instant file. 
Instant=[==>005__compaction__INFLIGHT][WARN ] 2020-12-12 15:14:53,395 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run[INFO] Tests run: 9, Failures: 0, Errors: 0, 
Skipped: 0, Time elapsed: 76.984 s - in 
org.apache.hudi.table.action.compact.TestAsyncCompaction[INFO] Running o
 rg.apache.hudi.table.action.compact.TestHoodieCompactor[WARN ] 2020-12-12 
15:14:54,314 org.apache.hudi.testutils.HoodieClientTestHarness  - Closing 
file-system instance used in previous test-run[WARN ] 2020-12-12 15:14:57,842 
org.apache.hudi.testutils.HoodieClientTestHarness  - Closing file-system 
instance used in previous test-run
   
   ## JIRA info
   
   - Link: https://issues.apache.org/jira/browse/HUDI-1454
   - Type: Task
   - Epic: https://issues.apache.org/jira/browse/HUDI-1042
   - Fix version(s):
     - 1.1.0


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to