Hi, all

We found some bugs in hive-0.12, but we could not wait for hive community fixing them.

We want to fix these bugs in our lab and build a new release which could be recognized by spark.

As we know, spark depends on a special release of hive, like:

|<dependency>
  <groupId>org.spark-project.hive</groupId>
  <artifactId>hive-metastore</artifactId>
  <version>${hive.version}</version>
</dependency>
|

The different between |org.spark-project.hive| and |org.apache.hive| was described by Patrick:

|There are two differences:

1. We publish hive with a shaded protobuf dependency to avoid
conflicts with some Hadoop versions.
2. We publish a proper hive-exec jar that only includes hive packages.
The upstream version of hive-exec bundles a bunch of other random
dependencies in it which makes it really hard for third-party projects
to use it.
|

Is there any document to guide us how to build the hive jars for spark?

Any help would be greatly appreciated.

​

Reply via email to