Hi all, I am trying to read data (using spark sql) via a hive metastore which has a column of type bigint. Underlying parquet data has int as the datatype for the same column. I am getting the following error while trying to read the data using spark sql -
java.lang.ClassCastException: org.apache.hadoop.io.IntWritable cannot be cast to org.apache.hadoop.io.LongWritable at org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableLongObjectInspector.get(WritableLongObjectInspector.java:36) at org.apache.spark.sql.hive.HadoopTableReader$$anonfun$14$$anonfun$apply$6.apply(TableReader.scala:418) ... I believe it is related to https://issues.apache.org/jira/browse/SPARK-17477. Any suggestions on how I can work around this issue? Spark version: 2.4.5 Regards, Naresh