Hi,

I am trying to read the multiple parquet files in sparksql. In one dir there
are two files, of which one is corrupted. While trying to read these files,
sparksql throws Exception for the corrupted file.

val newDataDF =
sqlContext.read.parquet("/data/testdir/data1.parquet","/data/testdir/corruptblock.0")
newDataDF.show 

throws Exception.

Is there any way to just skip the file having corrupted block/footer and
just read the file/files which are proper? 

Thanks



--
View this message in context: 
http://apache-spark-developers-list.1001551.n3.nabble.com/Skip-Corrupted-Parquet-blocks-footer-tp20418.html
Sent from the Apache Spark Developers List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to