Github user dongjoon-hyun commented on the issue:

    https://github.com/apache/spark/pull/19235
  
    @gatorsmile and @vanzin .
    
    I'm comparing with ORC now. Previously, ORC fails with another reason. I'll 
make another PR for that. I found that #14471 is enough for ORC.
    
    In case of ORC, ORC itself handles truncations on write. The padding is 
handled by Hive side `HiveCharWritable` via 
[HiveBaseChar.java](https://github.com/apache/hive/blob/master/common/src/java/org/apache/hadoop/hive/common/type/HiveBaseChar.java#L57)
 on read. In case of Parquet, I guess Parquet is the same, but there is no such 
a padding logic like HiveCharWritable in Spark.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to