zuston opened a new issue, #198:
URL: https://github.com/apache/incubator-uniffle/issues/198

   I found some tasks of spark jobs will throw the exceptions that the 
inconsistent blocks number. The stacktrace is as follows
   
   ```
   22/09/03 15:29:21 ERROR Executor: Exception in task 330.0 in stage 9.0 (TID 
59001)
   org.apache.uniffle.common.exception.RssException: Blocks read inconsistent: 
expected 30000 blocks, actual 15636 blocks
        at 
org.apache.uniffle.client.impl.ShuffleReadClientImpl.checkProcessedBlockIds(ShuffleReadClientImpl.java:215)
        at 
org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:135)
        at 
org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
   ```
   
   I didn't find any error/warn log in shuffle server which stored the 
corresponding partition data. 
   
   We dont set any replica config and directly use the MEMORY_LOCALFILE 
storageType. Does this exception caused by the disk error?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: dev-unsubscr...@uniffle.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to