[ 
https://issues.apache.org/jira/browse/SPARK-27913?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17034875#comment-17034875
 ] 

Giri commented on SPARK-27913:
------------------------------

This issue doesn't exist in  *spark spark-3.0.0-preview2 and also in spark 2.3* 
  Will this fix be ported to 2.4.x branch?

 

It appears that issue is realated to spark not using the schema from the 
metastore but from the ORC files and this causes the schema mismatch and out of 
bound exception when  OrcDeserializer accesses the field that doesn't exist in 
the file.

 

I see logs like this:

 

20/02/11 14:30:38 INFO RecordReaderImpl: Reader schema not provided -- using 
file schema struct<ar:struct<f1:int>>
20/02/11 14:30:38 INFO RecordReaderImpl: Reader schema not provided -- using 
file schema struct<a:struct<f1:int,f2:int>>

 

> Spark SQL's native ORC reader implements its own schema evolution
> -----------------------------------------------------------------
>
>                 Key: SPARK-27913
>                 URL: https://issues.apache.org/jira/browse/SPARK-27913
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 2.3.3
>            Reporter: Owen O'Malley
>            Priority: Major
>
> ORC's reader handles a wide range of schema evolution, but the Spark SQL 
> native ORC bindings do not provide the desired schema to the ORC reader. This 
> causes a regression when moving spark.sql.orc.impl from 'hive' to 'native'.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to