Thanks for reporting. As a workaround you should be able to SET spark.sql.hive.convertMetastoreParquet=false, but I'm going to try to fix this before the next RC.
On Wed, Dec 3, 2014 at 7:09 AM, Yana Kadiyska <yana.kadiy...@gmail.com> wrote: > Thanks Michael, you are correct. > > I also opened https://issues.apache.org/jira/browse/SPARK-4702 -- if > someone can comment on why this might be happening that would be great. > This would be a blocker to me using 1.2 and it used to work so I'm a bit > puzzled. I was hoping that it's again a result of the default profile > switch but it didn't seem to be the case > > (ps. please advise if this is more user-list appropriate. I'm posting to > dev as it's an RC) > > On Tue, Dec 2, 2014 at 8:37 PM, Michael Armbrust <mich...@databricks.com> > wrote: > >> In Hive 13 (which is the default for Spark 1.2), parquet is included and >> thus we no longer include the Hive parquet bundle. You can now use the >> included >> ParquetSerDe: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe >> >> If you want to compile Spark 1.2 with Hive 12 instead you can pass >> -Phive-0.12.0 and parquet.hive.serde.ParquetHiveSerDe will be included as >> before. >> >> Michael >> >> On Tue, Dec 2, 2014 at 9:31 AM, Yana Kadiyska <yana.kadiy...@gmail.com> >> wrote: >> >>> Apologies if people get this more than once -- I sent mail to dev@spark >>> last night and don't see it in the archives. Trying the incubator list >>> now...wanted to make sure it doesn't get lost in case it's a bug... >>> >>> ---------- Forwarded message ---------- >>> From: Yana Kadiyska <yana.kadiy...@gmail.com> >>> Date: Mon, Dec 1, 2014 at 8:10 PM >>> Subject: [Thrift,1.2 RC] what happened to >>> parquet.hive.serde.ParquetHiveSerDe >>> To: dev@spark.apache.org >>> >>> >>> Hi all, apologies if this is not a question for the dev list -- figured >>> User list might not be appropriate since I'm having trouble with the RC >>> tag. >>> >>> I just tried deploying the RC and running ThriftServer. I see the >>> following >>> error: >>> >>> 14/12/01 21:31:42 ERROR UserGroupInformation: PriviledgedActionException >>> as:anonymous (auth:SIMPLE) >>> cause:org.apache.hive.service.cli.HiveSQLException: >>> java.lang.RuntimeException: >>> MetaException(message:java.lang.ClassNotFoundException Class >>> parquet.hive.serde.ParquetHiveSerDe not found) >>> 14/12/01 21:31:42 WARN ThriftCLIService: Error executing statement: >>> org.apache.hive.service.cli.HiveSQLException: java.lang.RuntimeException: >>> MetaException(message:java.lang.ClassNotFoundException Class >>> parquet.hive.serde.ParquetHiveSerDe not found) >>> at >>> >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(Shim13.scala:192) >>> at >>> >>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:231) >>> at >>> >>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:212) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>> at >>> >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:606) >>> at >>> >>> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79) >>> at >>> >>> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37) >>> at >>> >>> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:415) >>> >>> >>> >>> I looked at a working installation that I have(build master a few weeks >>> ago) and this class used to be included in spark-assembly: >>> >>> ls *.jar|xargs grep parquet.hive.serde.ParquetHiveSerDe >>> Binary file spark-assembly-1.2.0-SNAPSHOT-hadoop2.0.0-mr1-cdh4.2.0.jar >>> matches >>> >>> but with the RC build it's not there? >>> >>> I tried both the prebuilt CDH drop and later manually built the tag with >>> the following command: >>> >>> ./make-distribution.sh --tgz -Phive -Dhadoop.version=2.0.0-mr1-cdh4.2.0 >>> -Phive-thriftserver >>> $JAVA_HOME/bin/jar -tvf spark-assembly-1.2.0-hadoop2.0.0-mr1-cdh4.2.0.jar >>> |grep parquet.hive.serde.ParquetHiveSerDe >>> >>> comes back empty... >>> >> >> >