This should fix it: https://github.com/apache/spark/pull/16080
On Wed, Nov 30, 2016 at 10:55 AM, Timur Shenkao <t...@timshenkao.su> wrote: > Hello, > > Yes, I used hiveContext, sqlContext, sparkSession from Java, Scala, > Python. > Via spark-shell, spark-submit, IDE (PyCharm, Intellij IDEA). > Everything is perfect because I have Hadoop cluster with configured & > tuned HIVE. > > The reason of Michael's error is usually misconfigured or absent HIVE. > Or may be absence of hive-site.xml in $SPARK_HOME/conf/ directory. > > On Wed, Nov 30, 2016 at 9:30 PM, Gourav Sengupta < > gourav.sengu...@gmail.com> wrote: > >> Hi Timur, >> >> did you use hiveContext or sqlContext or the spark way mentioned in the >> http://spark.apache.org/docs/latest/sql-programming-guide.html? >> >> >> Regards, >> Gourav Sengupta >> >> On Wed, Nov 30, 2016 at 5:35 PM, Yin Huai <yh...@databricks.com> wrote: >> >>> Hello Michael, >>> >>> Thank you for reporting this issue. It will be fixed by >>> https://github.com/apache/spark/pull/16080. >>> >>> Thanks, >>> >>> Yin >>> >>> On Tue, Nov 29, 2016 at 11:34 PM, Timur Shenkao <t...@timshenkao.su> >>> wrote: >>> >>>> Hi! >>>> >>>> Do you have real HIVE installation? >>>> Have you built Spark 2.1 & Spark 2.0 with HIVE support ( -Phive >>>> -Phive-thriftserver ) ? >>>> >>>> It seems that you use "default" Spark's HIVE 1.2.1. Your metadata is >>>> stored in local Derby DB which is visible to concrete Spark installation >>>> but not for all. >>>> >>>> On Wed, Nov 30, 2016 at 4:51 AM, Michael Allman <mich...@videoamp.com> >>>> wrote: >>>> >>>>> This is not an issue with all tables created in Spark 2.1, though I'm >>>>> not sure why some work and some do not. I have found that a table created >>>>> as such >>>>> >>>>> sql("create table test stored as parquet as select 1") >>>>> >>>>> in Spark 2.1 cannot be read in previous versions of Spark. >>>>> >>>>> Michael >>>>> >>>>> >>>>> > On Nov 29, 2016, at 5:15 PM, Michael Allman <mich...@videoamp.com> >>>>> wrote: >>>>> > >>>>> > Hello, >>>>> > >>>>> > When I try to read from a Hive table created by Spark 2.1 in Spark >>>>> 2.0 or earlier, I get an error: >>>>> > >>>>> > java.lang.ClassNotFoundException: Failed to load class for data >>>>> source: hive. >>>>> > >>>>> > Is there a way to get previous versions of Spark to read tables >>>>> written with Spark 2.1? >>>>> > >>>>> > Cheers, >>>>> > >>>>> > Michael >>>>> >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>>> >>>>> >>>> >>> >> >