In 1.5.x whenever I try to create a HiveContext from SparkContext I get following error. Please note that I'm not running any Hadoop/Hive server in my cluster. I'm only running Spark.
I never faced HiveContext creation problem like this previously in 1.4.x. Is it now a requirement in 1.5.x that to create HIveContext Hive Server should be running ? Regards, Sourav ---------- Forwarded message ---------- From: <ross.cramb...@thomsonreuters.com> Date: Fri, Dec 11, 2015 at 11:39 AM Subject: Re: Window function in Spark SQL To: sourav.mazumde...@gmail.com I’m not familiar with that issue, I wasn’t able to reproduce in my environment - might want to copy that to the Spark user list. Sorry! On Dec 11, 2015, at 1:37 PM, Sourav Mazumder <sourav.mazumde...@gmail.com> wrote: Hi Ross, Thanks for your answer. In 1.5.x whenever I try to create a HiveContext from SparkContext I get following error. Please note that I'm not running any Hadoop/Hive server in my cluster. I'm only running Spark. I never faced HiveContext creation problem like this previously Regards, Sourav java.lang.RuntimeException: java.lang.RuntimeException: The root scratch dir: /t mp/hive on HDFS should be writable. Current permissions are: rwx------ at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.jav a:522) at org.apache.spark.sql.hive.client.ClientWrapper.<init>(ClientWrapper.s cala:171) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstruct orAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingC onstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:422) at org.apache.spark.sql.hive.client.IsolatedClientLoader.liftedTree1$1(I solatedClientLoader.scala:183) at org.apache.spark.sql.hive.client.IsolatedClientLoader.<init>(Isolated ClientLoader.scala:179) at org.apache.spark.sql.hive.HiveContext.metadataHive$lzycompute(HiveCon text.scala:226) at org.apache.spark.sql.hive.HiveContext.metadataHive(HiveContext.scala: 185) at org.apache.spark.sql.hive.HiveContext.setConf(HiveContext.scala:392) at org.apache.spark.sql.hive.HiveContext.defaultOverrides(HiveContext.sc ala:174) at org.apache.spark.sql.hive.HiveContext.<init>(HiveContext.scala:177) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:15) at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:20) at $iwC$$iwC$$iwC$$iwC.<init>(<console>:22) at $iwC$$iwC$$iwC.<init>(<console>:24) at $iwC$$iwC.<init>(<console>:26) at $iwC.<init>(<console>:28) at <init>(<console>:30) at .<init>(<console>:34) at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at $print(<console>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) On Fri, Dec 11, 2015 at 10:10 AM, <ross.cramb...@thomsonreuters.com> wrote: > Hey Sourav, > Window functions require using a HiveContext rather than the default > SQLContext. See here: > http://spark.apache.org/docs/latest/sql-programming-guide.html#starting-point-sqlcontext > <https://urldefense.proofpoint.com/v2/url?u=http-3A__spark.apache.org_docs_latest_sql-2Dprogramming-2Dguide.html-23starting-2Dpoint-2Dsqlcontext&d=AwMFaQ&c=4ZIZThykDLcoWk-GVjSLm9hvvvzvGv0FLoWSRuCSs5Q&r=DJcC0Gr3B6BfuPcycQUvAi5ueGCorF1rF8_kDa-hAYg&m=3CetCpYEuIP9Jnpnh_Rsg7OdguGE6FK3u0ByvvHvwpA&s=gJ-XUDEXWy54su9cV-7f-AmnPmbwMxOzsEK9Q_K9qu4&e=> > > HiveContext provides all the same functionality of SQLContext, as well as > extra features like Window functions. > > - Ross > > On Dec 11, 2015, at 12:59 PM, Sourav Mazumder <sourav.mazumde...@gmail.com> > wrote: > > Hi, > > Spark SQL documentation says that it complies with Hive 1.2.1 APIs and > supports Window functions. I'm using Spark 1.5.0. > > However, when I try to execute something like below I get an error > > val lol5 = sqlContext.sql("select ky, lead(ky, 5, 0) over (order by ky > rows 5 following) from lolt") > > java.lang.RuntimeException: [1.32] failure: ``union'' expected but `(' > found select ky, lead(ky, 5, 0) over (order by ky rows 5 following) from > lolt ^ at scala.sys.package$.error(package.scala:27) at > org.apache.spark.sql.catalyst.AbstractSparkSQLParser.parse(AbstractSparkSQLParser.scala:36) > at > org.apache.spark.sql.catalyst.DefaultParserDialect.parse(ParserDialect.scala:67) > at org.apache.spark.sql.SQLContext$$anonfun$3.apply(SQLContext.scala:169) > at org.apache.spark.sql.SQLContext$$anonfun$3.apply(SQLContext.scala:169) > at > org.apache.spark.sql.SparkSQLParser$$anonfun$org$apache$spark$sql$SparkSQLParser$$others$1.apply(SparkSQLParser.scala:115) > at > org.apache.spark.sql.SparkSQLParser$$anonfun$org$apache$spark$sql$SparkSQLParser$$others$1.apply(SparkSQLParser.scala:114) > at scala.util.parsing.combinator.Parsers$Success.map(Parsers.scala:136) at > scala.util.parsing.combinator.Parsers$Success.map(Parsers.scala:135) at > scala.util.parsing.combinator.Parsers$Parser$$anonfun$map$1.apply(Parsers.scala:242) > at > scala.util.parsing.combinator.Parsers$Parser$$anonfun$map$1.apply(Parsers.scala:242) > at scala.util.parsing.combinator.Parsers$$anon$3.apply(Parsers.scala:222) > at > scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1$$anonfun$apply$2.apply(Parsers.scala:254) > at > scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1$$anonfun$apply$2.apply(Parsers.scala:254) > at scala.util.parsing.combinator.Parsers$Failure.append(Parsers.scala:202) > at > scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1.apply(Parsers.scala:254) > at > scala.util.parsing.combinator.Parsers$Parser$$anonfun$append$1.apply(Parsers.scala:254) > at scala.util.parsing.combinator.Parsers$$anon$3.apply(Parsers.scala:222) > at > scala.util.parsing.combinator.Parsers$$anon$2$$anonfun$apply$14.apply(Parsers.scala:891) > at > scala.util.parsing.combinator.Parsers$$anon$2$$anonfun$apply$14.apply(Parsers.scala:891) > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) at > scala.util.parsing.combinator.Parsers$$anon$2.apply(Parsers.scala:890) at > scala.util.parsing.combinator.PackratParsers$$anon$1.apply(PackratParsers.scala:110) > at > org.apache.spark.sql.catalyst.AbstractSparkSQLParser.parse(AbstractSparkSQLParser.scala:34) > at org.apache.spark.sql.SQLContext$$anonfun$2.apply(SQLContext.scala:166) > at org.apache.spark.sql.SQLContext$$anonfun$2.apply(SQLContext.scala:166) > at > org.apache.spark.sql.execution.datasources.DDLParser.parse(DDLParser.scala:42) > at org.apache.spark.sql.SQLContext.parseSql(SQLContext.scala:189) at > org.apache.spark.sql.SQLContext.sql(SQLContext.scala:719) at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:63) > at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:68) > at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:70) > at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:72) > at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:74) > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:76) > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:78) at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:80) at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:82) at > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:84) at > $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:86) at > $iwC$$iwC$$iwC$$iwC.<init>(<console>:88) at > $iwC$$iwC$$iwC.<init>(<console>:90) at $iwC$$iwC.<init>(<console>:92) at > $iwC.<init>(<console>:94) at <init>(<console>:96) at .<init>(<console>:100) > at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at > $print(<console>) > > Regards, > Sourav > > >