[jira] [Assigned] (SPARK-34006) [spark.sql.hive.convertMetastoreOrc]This parameter can solve orc format table insert overwrite read table, it should be stated in the document

2021-01-04 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-34006?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-34006:


Assignee: (was: Apache Spark)

> [spark.sql.hive.convertMetastoreOrc]This parameter can solve orc format table 
> insert overwrite read table, it should be stated in the document
> --
>
> Key: SPARK-34006
> URL: https://issues.apache.org/jira/browse/SPARK-34006
> Project: Spark
>  Issue Type: Bug
>  Components: docs
>Affects Versions: 3.0.1
>Reporter: hao
>Priority: Major
>
> This parameter can solve orc format table insert overwrite read table, it 
> should be stated in the document



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-34006) [spark.sql.hive.convertMetastoreOrc]This parameter can solve orc format table insert overwrite read table, it should be stated in the document

2021-01-04 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-34006?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-34006:


Assignee: Apache Spark

> [spark.sql.hive.convertMetastoreOrc]This parameter can solve orc format table 
> insert overwrite read table, it should be stated in the document
> --
>
> Key: SPARK-34006
> URL: https://issues.apache.org/jira/browse/SPARK-34006
> Project: Spark
>  Issue Type: Bug
>  Components: docs
>Affects Versions: 3.0.1
>Reporter: hao
>Assignee: Apache Spark
>Priority: Major
>
> This parameter can solve orc format table insert overwrite read table, it 
> should be stated in the document



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org