In 1.5.x whenever I try to create a HiveContext from SparkContext I get
following error. Please note that I'm not running any Hadoop/Hive server in
my cluster. I'm only running Spark.

I never faced HiveContext creation problem like this previously in 1.4.x.

Is it now a requirement in 1.5.x that to create HIveContext Hive Server
should be running ?

Regards,
Sourav


---------- Forwarded message ----------
From: <ross.cramb...@thomsonreuters.com>
Date: Fri, Dec 11, 2015 at 11:39 AM
Subject: Re: Window function in Spark SQL
To: sourav.mazumde...@gmail.com


I’m not familiar with that issue, I wasn’t able to reproduce in my
environment - might want to copy that to the Spark user list. Sorry!

On Dec 11, 2015, at 1:37 PM, Sourav Mazumder <sourav.mazumde...@gmail.com>
wrote:

Hi Ross,

Thanks for your answer.

In 1.5.x whenever I try to create a HiveContext from SparkContext I get
following error. Please note that I'm not running any Hadoop/Hive server in
my cluster. I'm only running Spark.

I never faced HiveContext creation problem like this previously

Regards,
Sourav

java.lang.RuntimeException: java.lang.RuntimeException: The root scratch
dir: /t
mp/hive on HDFS should be writable. Current permissions are: rwx------
        at
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.jav
a:522)
        at
org.apache.spark.sql.hive.client.ClientWrapper.<init>(ClientWrapper.s
cala:171)
        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)

        at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstruct
orAccessorImpl.java:62)
        at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingC
onstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
        at
org.apache.spark.sql.hive.client.IsolatedClientLoader.liftedTree1$1(I
solatedClientLoader.scala:183)
        at
org.apache.spark.sql.hive.client.IsolatedClientLoader.<init>(Isolated
ClientLoader.scala:179)
        at
org.apache.spark.sql.hive.HiveContext.metadataHive$lzycompute(HiveCon
text.scala:226)
        at
org.apache.spark.sql.hive.HiveContext.metadataHive(HiveContext.scala:
185)
        at
org.apache.spark.sql.hive.HiveContext.setConf(HiveContext.scala:392)
        at
org.apache.spark.sql.hive.HiveContext.defaultOverrides(HiveContext.sc
ala:174)
        at
org.apache.spark.sql.hive.HiveContext.<init>(HiveContext.scala:177)
        at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:15)
        at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:20)
        at $iwC$$iwC$$iwC$$iwC.<init>(<console>:22)
        at $iwC$$iwC$$iwC.<init>(<console>:24)
        at $iwC$$iwC.<init>(<console>:26)
        at $iwC.<init>(<console>:28)
        at <init>(<console>:30)
        at .<init>(<console>:34)
        at .<clinit>(<console>)
        at .<init>(<console>:7)
        at .<clinit>(<console>)
        at $print(<console>)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

On Fri, Dec 11, 2015 at 10:10 AM, <ross.cramb...@thomsonreuters.com> wrote:

> Hey Sourav,
> Window functions require using a HiveContext rather than the default
> SQLContext. See here:
> http://spark.apache.org/docs/latest/sql-programming-guide.html#starting-point-sqlcontext
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__spark.apache.org_docs_latest_sql-2Dprogramming-2Dguide.html-23starting-2Dpoint-2Dsqlcontext&d=AwMFaQ&c=4ZIZThykDLcoWk-GVjSLm9hvvvzvGv0FLoWSRuCSs5Q&r=DJcC0Gr3B6BfuPcycQUvAi5ueGCorF1rF8_kDa-hAYg&m=3CetCpYEuIP9Jnpnh_Rsg7OdguGE6FK3u0ByvvHvwpA&s=gJ-XUDEXWy54su9cV-7f-AmnPmbwMxOzsEK9Q_K9qu4&e=>
>
> HiveContext provides all the same functionality of SQLContext, as well as
> extra features like Window functions.
>
> - Ross
>
> On Dec 11, 2015, at 12:59 PM, Sourav Mazumder <sourav.mazumde...@gmail.com>
> wrote:
>
> Hi,
>
> Spark SQL documentation says that it complies with Hive 1.2.1 APIs and
> supports Window functions. I'm using Spark 1.5.0.
>
> However, when I try to execute something like below I get an error
>
> val lol5 = sqlContext.sql("select ky, lead(ky, 5, 0) over (order by ky
> rows 5 following) from lolt")
>
> java.lang.RuntimeException: [1.32] failure: ``union'' expected but `('
> found select ky, lead(ky, 5, 0) over (order by ky rows 5 following) from
> lolt ^ at scala.sys.package$.error(package.scala:27) at
> org.apache.spark.sql.catalyst.AbstractSparkSQLParser.parse(AbstractSparkSQLParser.scala:36)
> at
> org.apache.spark.sql.catalyst.DefaultParserDialect.parse(ParserDialect.scala:67)
> at org.apache.spark.sql.SQLContext$$anonfun$3.apply(SQLContext.scala:169)
> at org.apache.spark.sql.SQLContext$$anonfun$3.apply(SQLContext.scala:169)
> at
> org.apache.spark.sql.SparkSQLParser$$anonfun$org$apache$spark$sql$SparkSQLParser$$others$1.apply(SparkSQLParser.scala:115)
> at
> org.apache.spark.sql.SparkSQLParser$$anonfun$org$apache$spark$sql$SparkSQLParser$$others$1.apply(SparkSQLParser.scala:114)
> at scala.util.parsing.combinator.Parsers$Success.map(Parsers.scala:136) at
> scala.util.parsing.combinator.Parsers$Success.map(Parsers.scala:135) at
> scala.util.parsing.combinator.Parsers$Parser$$anonfun$map$1.apply(Parsers.scala:242)
> at
> scala.util.parsing.combinator.Parsers$Parser$$anonfun$map$1.apply(Parsers.scala:242)
> at scala.util.parsing.combinator.Parsers$$anon$3.apply(Parsers.scala:222)
> at
> scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1$$anonfun$apply$2.apply(Parsers.scala:254)
> at
> scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1$$anonfun$apply$2.apply(Parsers.scala:254)
> at scala.util.parsing.combinator.Parsers$Failure.append(Parsers.scala:202)
> at
> scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1.apply(Parsers.scala:254)
> at
> scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1.apply(Parsers.scala:254)
> at scala.util.parsing.combinator.Parsers$$anon$3.apply(Parsers.scala:222)
> at
> scala.util.parsing.combinator.Parsers$$anon$2$$anonfun$apply$14.apply(Parsers.scala:891)
> at
> scala.util.parsing.combinator.Parsers$$anon$2$$anonfun$apply$14.apply(Parsers.scala:891)
> at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) at
> scala.util.parsing.combinator.Parsers$$anon$2.apply(Parsers.scala:890) at
> scala.util.parsing.combinator.PackratParsers$$anon$1.apply(PackratParsers.scala:110)
> at
> org.apache.spark.sql.catalyst.AbstractSparkSQLParser.parse(AbstractSparkSQLParser.scala:34)
> at org.apache.spark.sql.SQLContext$$anonfun$2.apply(SQLContext.scala:166)
> at org.apache.spark.sql.SQLContext$$anonfun$2.apply(SQLContext.scala:166)
> at
> org.apache.spark.sql.execution.datasources.DDLParser.parse(DDLParser.scala:42)
> at org.apache.spark.sql.SQLContext.parseSql(SQLContext.scala:189) at
> org.apache.spark.sql.SQLContext.sql(SQLContext.scala:719) at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:63)
> at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:68)
> at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:70)
> at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:72)
> at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:74)
> at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:76)
> at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:78) at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:80) at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:82) at
> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:84) at
> $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:86) at
> $iwC$$iwC$$iwC$$iwC.<init>(<console>:88) at
> $iwC$$iwC$$iwC.<init>(<console>:90) at $iwC$$iwC.<init>(<console>:92) at
> $iwC.<init>(<console>:94) at <init>(<console>:96) at .<init>(<console>:100)
> at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at
> $print(<console>)
>
> Regards,
> Sourav
>
>
>

Reply via email to