Thank you Hao for the confirmation!

I filed two jiras as follows-
https://issues.apache.org/jira/browse/SPARK-7850 (removing hive-0.12.0
profile from pom)
https://issues.apache.org/jira/browse/SPARK-7851 (thrift error with hive
metastore 0.12)


On Sun, May 24, 2015 at 8:18 PM, Cheng, Hao <hao.ch...@intel.com> wrote:

>  Thanks for reporting this.
>
>
>
> We intend to support the multiple metastore versions in a single
> build(hive-0.13.1) by introducing the IsolatedClientLoader, but probably
> you’re hitting the bug, please file a jira issue for this.
>
>
>
> I will keep investigating on this also.
>
>
>
> Hao
>
>
>
>
>
> *From:* Mark Hamstra [mailto:m...@clearstorydata.com]
> *Sent:* Sunday, May 24, 2015 9:06 PM
> *To:* Cheolsoo Park
> *Cc:* u...@spark.apache.org; dev@spark.apache.org
> *Subject:* Re: SparkSQL errors in 1.4 rc when using with Hive 0.12
> metastore
>
>
>
> This discussion belongs on the dev list.  Please post any replies there.
>
>
>
> On Sat, May 23, 2015 at 10:19 PM, Cheolsoo Park <piaozhe...@gmail.com>
> wrote:
>
>  Hi,
>
>
>
> I've been testing SparkSQL in 1.4 rc and found two issues. I wanted to
> confirm whether these are bugs or not before opening a jira.
>
>
> *1)* I can no longer compile SparkSQL with -Phive-0.12.0. I noticed that
> in 1.4, IsolatedClientLoader is introduced, and different versions of Hive
> metastore jars can be loaded at runtime. But instead, SparkSQL no longer
> compiles with Hive 0.12.0.
>
>
>
> My question is, is this intended? If so, shouldn't the hive-0.12.0 profile
> in POM be removed?
>
>
>
> *2)* After compiling SparkSQL with -Phive-0.13.1, I ran into my 2nd
> problem. Since I have Hive 0.12 metastore in production, I have to use it
> for now. But even if I set "spark.sql.hive.metastore.version" and
> "spark.sql.hive.metastore.jars", SparkSQL cli throws an error as follows-
>
>
>
> 15/05/24 05:03:29 WARN RetryingMetaStoreClient: MetaStoreClient lost
> connection. Attempting to reconnect.
>
> org.apache.thrift.TApplicationException: Invalid method name:
> 'get_functions'
>
> at
> org.apache.thrift.TApplicationException.read(TApplicationException.java:108)
>
> at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:71)
>
> at
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_functions(ThriftHiveMetastore.java:2886)
>
> at
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_functions(ThriftHiveMetastore.java:2872)
>
> at
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getFunctions(HiveMetaStoreClient.java:1727)
>
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>
> at java.lang.reflect.Method.invoke(Method.java:606)
>
> at
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:89)
>
> at com.sun.proxy.$Proxy12.getFunctions(Unknown Source)
>
> at org.apache.hadoop.hive.ql.metadata.Hive.getFunctions(Hive.java:2670)
>
> at
> org.apache.hadoop.hive.ql.exec.FunctionRegistry.getFunctionNames(FunctionRegistry.java:674)
>
> at
> org.apache.hadoop.hive.ql.exec.FunctionRegistry.getFunctionNames(FunctionRegistry.java:662)
>
> at
> org.apache.hadoop.hive.cli.CliDriver.getCommandCompletor(CliDriver.java:540)
>
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:175)
>
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
>
>
>
> What's happening is that when SparkSQL Cli starts up, it tries to fetch
> permanent udfs from Hive metastore (due to HIVE-6330
> <https://issues.apache.org/jira/browse/HIVE-6330>, which was introduced
> in Hive 0.13). But then, it ends up invoking an incompatible thrift
> function that doesn't exist in Hive 0.12. To work around this error, I have
> to comment out the following line of code for now-
>
> https://goo.gl/wcfnH1
>
>
>
> My question is, is SparkSQL that is compiled against Hive 0.13 supposed to
> work with Hive 0.12 metastore (by setting
> "spark.sql.hive.metastore.version" and "spark.sql.hive.metastore.jars")? It
> only works if I comment out the above line of code.
>
>
>
> Thanks,
>
> Cheolsoo
>
>
>

Reply via email to