Github user dongjoon-hyun commented on the issue: https://github.com/apache/spark/pull/19235 @gatorsmile and @vanzin . I'm comparing with ORC now. Previously, ORC fails with another reason. I'll make another PR for that. I found that #14471 is enough for ORC. In case of ORC, ORC itself handles truncations on write. The padding is handled by Hive side `HiveCharWritable` via [HiveBaseChar.java](https://github.com/apache/hive/blob/master/common/src/java/org/apache/hadoop/hive/common/type/HiveBaseChar.java#L57) on read. In case of Parquet, I guess Parquet is the same, but there is no such a padding logic like HiveCharWritable in Spark.
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org