Sagar Sumit created HUDI-2526:
---------------------------------

             Summary: Make spark.sql.parquet.writeLegacyFormat configurable
                 Key: HUDI-2526
                 URL: https://issues.apache.org/jira/browse/HUDI-2526
             Project: Apache Hudi
          Issue Type: Improvement
            Reporter: Sagar Sumit
            Assignee: sivabalan narayanan


>From the community, 

"I am observing that HUDI bulk inser in 0.9.0 version is not honoring
spark.sql.parquet.writeLegacyFormat=true
config. Can you suggest way to set this config.
Reason to use this config:
Current Bulk insert use spark dataframe writer and don't do avro conversion. 
The decimal columns in my DF are written as INT32 type in parquet.
The upsert functionality which uses avro conversion is generating Fixed Length 
byte array for decimal types which is failing with datatype mismatch." 

The main reason is that the [config is 
hardcoded|https://github.com/apache/hudi/blob/46808dcb1fe22491326a9e831dd4dde4c70796fb/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/io/storage/row/HoodieRowParquetWriteSupport.java#L48].
 We can make it configurable.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to