[ 
https://issues.apache.org/jira/browse/SPARK-20202?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15953304#comment-15953304
 ] 

Sean Owen commented on SPARK-20202:
-----------------------------------

Agree. I think the logic was that Spark had released its own source/binary 
version of Hive, and then used that in Spark. I don't think anybody believes 
that's a good solution in the long term; it was a work-around for hive-exec's 
packaging IIRC. Once whatever that is is resolved this can go away, but I defer 
to those who know the issue better on the details.

What I'm not clear on is whether the current org.spark-hive situation is 
streeetching the source/binary policy so far that it breaks, enough that no 
more releases can happen without it. Best to make it go away ASAP anyway. But I 
don't know if changes in Hive 2.5 help integration with Hive 1.x. It may 
require either temporarily blessing the fork, or more jar surgery to un-uberize 
the hive-exec jar or something.

> Remove references to org.spark-project.hive
> -------------------------------------------
>
>                 Key: SPARK-20202
>                 URL: https://issues.apache.org/jira/browse/SPARK-20202
>             Project: Spark
>          Issue Type: Bug
>          Components: Build, SQL
>    Affects Versions: 1.6.4, 2.0.3, 2.1.1
>            Reporter: Owen O'Malley
>            Priority: Critical
>
> Spark can't continue to depend on their fork of Hive and must move to 
> standard Hive versions.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to