Jinhua Fu created SPARK-21786:
---------------------------------

             Summary: The 'spark.sql.parquet.compression.codec' configuration 
doesn't take effect on tables with partition field(s)
                 Key: SPARK-21786
                 URL: https://issues.apache.org/jira/browse/SPARK-21786
             Project: Spark
          Issue Type: Bug
          Components: SQL
    Affects Versions: 2.2.0
            Reporter: Jinhua Fu


For tables created like below,  'spark.sql.parquet.compression.codec' doesn't 
take any effect when insert data. And because the default compression codec is 
'uncompressed', if I want to change the compression codec, I have to change it 
by 'set parquet.compression=xxxx'.

Contrast,tables without any partition field will work normal with 
'spark.sql.parquet.compression.codec',and the default compression codec is 
'snappy', but it seems 'parquet.compression' no longer in effect.

Should we use the ‘spark.sql.parquet.compression.codec’ configuration uniformly?

----------------------------------------------------------------------------------------------------
CREATE TABLE Test_Parquet(provincecode int, citycode int, districtcode int)
PARTITIONED BY (p_provincecode int)
STORED AS PARQUET;

INSERT OVERWRITE TABLE Test_Parquet select * from TableB;



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to